[ 498.239411] env[61906]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=61906) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 498.239731] env[61906]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=61906) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 498.239840] env[61906]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=61906) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 498.240180] env[61906]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 498.333085] env[61906]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=61906) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 498.342871] env[61906]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=61906) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 498.951991] env[61906]: INFO nova.virt.driver [None req-f5e415a1-6a17-41d6-aea8-130c43f3bd49 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 499.021295] env[61906]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 499.021535] env[61906]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 499.021535] env[61906]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=61906) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 502.145789] env[61906]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-2839c73a-0780-4d0d-8cb5-e1767a701747 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.161594] env[61906]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=61906) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 502.161816] env[61906]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-218bdf16-f7e7-4d9b-bb24-9abac437569a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.189579] env[61906]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 020e2. [ 502.189714] env[61906]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.168s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 502.190302] env[61906]: INFO nova.virt.vmwareapi.driver [None req-f5e415a1-6a17-41d6-aea8-130c43f3bd49 None None] VMware vCenter version: 7.0.3 [ 502.193659] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a616a330-6379-4c2e-8d15-9b2c55f9d89c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.215546] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a5e222d-9fef-4b6f-acf2-8a01cdf17c5a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.221699] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4286012d-5f61-4465-8e1e-4e5d794bd629 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.228301] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1dada73-cc4d-4e7a-a0b6-0919011aad5d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.241345] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a69d885a-2ff6-4f22-99c4-8523487202ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.247058] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c93946fe-2660-4bd0-be29-1c70baf2ccf2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.276433] env[61906]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-a1ebe6b0-0bb7-47d9-aa08-732a8dac92d6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.281260] env[61906]: DEBUG nova.virt.vmwareapi.driver [None req-f5e415a1-6a17-41d6-aea8-130c43f3bd49 None None] Extension org.openstack.compute already exists. {{(pid=61906) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:226}} [ 502.283814] env[61906]: INFO nova.compute.provider_config [None req-f5e415a1-6a17-41d6-aea8-130c43f3bd49 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 502.787058] env[61906]: DEBUG nova.context [None req-f5e415a1-6a17-41d6-aea8-130c43f3bd49 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),e9083091-9e1e-409c-b3f7-134c81f4733b(cell1) {{(pid=61906) load_cells /opt/stack/nova/nova/context.py:464}} [ 502.789200] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 502.789422] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 502.790092] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 502.790531] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] Acquiring lock "e9083091-9e1e-409c-b3f7-134c81f4733b" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 502.790734] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] Lock "e9083091-9e1e-409c-b3f7-134c81f4733b" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 502.791746] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] Lock "e9083091-9e1e-409c-b3f7-134c81f4733b" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 502.811442] env[61906]: INFO dbcounter [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] Registered counter for database nova_cell0 [ 502.819277] env[61906]: INFO dbcounter [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] Registered counter for database nova_cell1 [ 502.822589] env[61906]: DEBUG oslo_db.sqlalchemy.engines [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61906) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 502.823184] env[61906]: DEBUG oslo_db.sqlalchemy.engines [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61906) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 502.827538] env[61906]: ERROR nova.db.main.api [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 502.827538] env[61906]: result = function(*args, **kwargs) [ 502.827538] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 502.827538] env[61906]: return func(*args, **kwargs) [ 502.827538] env[61906]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 502.827538] env[61906]: result = fn(*args, **kwargs) [ 502.827538] env[61906]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 502.827538] env[61906]: return f(*args, **kwargs) [ 502.827538] env[61906]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 502.827538] env[61906]: return db.service_get_minimum_version(context, binaries) [ 502.827538] env[61906]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 502.827538] env[61906]: _check_db_access() [ 502.827538] env[61906]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 502.827538] env[61906]: stacktrace = ''.join(traceback.format_stack()) [ 502.827538] env[61906]: [ 502.828496] env[61906]: ERROR nova.db.main.api [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 502.828496] env[61906]: result = function(*args, **kwargs) [ 502.828496] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 502.828496] env[61906]: return func(*args, **kwargs) [ 502.828496] env[61906]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 502.828496] env[61906]: result = fn(*args, **kwargs) [ 502.828496] env[61906]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 502.828496] env[61906]: return f(*args, **kwargs) [ 502.828496] env[61906]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 502.828496] env[61906]: return db.service_get_minimum_version(context, binaries) [ 502.828496] env[61906]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 502.828496] env[61906]: _check_db_access() [ 502.828496] env[61906]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 502.828496] env[61906]: stacktrace = ''.join(traceback.format_stack()) [ 502.828496] env[61906]: [ 502.828912] env[61906]: WARNING nova.objects.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 502.829016] env[61906]: WARNING nova.objects.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] Failed to get minimum service version for cell e9083091-9e1e-409c-b3f7-134c81f4733b [ 502.829438] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] Acquiring lock "singleton_lock" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 502.829600] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] Acquired lock "singleton_lock" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 502.829839] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] Releasing lock "singleton_lock" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 502.830169] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] Full set of CONF: {{(pid=61906) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 502.830319] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ******************************************************************************** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 502.830443] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] Configuration options gathered from: {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 502.830584] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 502.830776] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 502.830903] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ================================================================================ {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 502.831122] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] allow_resize_to_same_host = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.831296] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] arq_binding_timeout = 300 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.831432] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] backdoor_port = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.831559] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] backdoor_socket = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.831720] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] block_device_allocate_retries = 60 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.831877] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] block_device_allocate_retries_interval = 3 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.832055] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cert = self.pem {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.832226] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.832394] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] compute_monitors = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.832561] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] config_dir = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.832726] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] config_drive_format = iso9660 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.832856] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.833034] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] config_source = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.833207] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] console_host = devstack {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.833373] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] control_exchange = nova {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.833528] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cpu_allocation_ratio = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.833686] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] daemon = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.833849] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] debug = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.834010] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] default_access_ip_network_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.834180] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] default_availability_zone = nova {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.834336] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] default_ephemeral_format = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.834493] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] default_green_pool_size = 1000 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.834724] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.834888] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] default_schedule_zone = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.835053] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] disk_allocation_ratio = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.835215] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] enable_new_services = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.835387] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] enabled_apis = ['osapi_compute'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.835548] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] enabled_ssl_apis = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.835706] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] flat_injected = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.835861] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] force_config_drive = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.836033] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] force_raw_images = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.836204] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] graceful_shutdown_timeout = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.836364] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] heal_instance_info_cache_interval = 60 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.836608] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] host = cpu-1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.836737] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.836896] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] initial_disk_allocation_ratio = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.837067] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] initial_ram_allocation_ratio = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.837280] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.837444] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] instance_build_timeout = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.837599] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] instance_delete_interval = 300 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.837763] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] instance_format = [instance: %(uuid)s] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.837927] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] instance_name_template = instance-%08x {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.838101] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] instance_usage_audit = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.838278] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] instance_usage_audit_period = month {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.838446] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.838614] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] instances_path = /opt/stack/data/nova/instances {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.838780] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] internal_service_availability_zone = internal {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.838936] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] key = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.839106] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] live_migration_retry_count = 30 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.839277] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] log_color = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.839440] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] log_config_append = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.839603] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.839762] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] log_dir = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.839919] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] log_file = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.840059] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] log_options = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.840222] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] log_rotate_interval = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.840389] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] log_rotate_interval_type = days {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.840552] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] log_rotation_type = none {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.840679] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.840805] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.840969] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.841144] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.841275] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.841435] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] long_rpc_timeout = 1800 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.841593] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] max_concurrent_builds = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.841750] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] max_concurrent_live_migrations = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.841908] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] max_concurrent_snapshots = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.842072] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] max_local_block_devices = 3 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.842235] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] max_logfile_count = 30 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.842391] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] max_logfile_size_mb = 200 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.842547] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] maximum_instance_delete_attempts = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.842712] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] metadata_listen = 0.0.0.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.842881] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] metadata_listen_port = 8775 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.843056] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] metadata_workers = 2 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.843218] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] migrate_max_retries = -1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.843384] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] mkisofs_cmd = genisoimage {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.843588] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] my_block_storage_ip = 10.180.1.21 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.843718] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] my_ip = 10.180.1.21 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.843877] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] network_allocate_retries = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.844065] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.844237] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] osapi_compute_listen = 0.0.0.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.844402] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] osapi_compute_listen_port = 8774 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.844567] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] osapi_compute_unique_server_name_scope = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.844732] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] osapi_compute_workers = 2 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.844892] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] password_length = 12 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.845062] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] periodic_enable = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.845225] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] periodic_fuzzy_delay = 60 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.845391] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] pointer_model = usbtablet {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.845555] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] preallocate_images = none {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.845711] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] publish_errors = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.845840] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] pybasedir = /opt/stack/nova {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.845996] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ram_allocation_ratio = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.846168] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] rate_limit_burst = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.846337] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] rate_limit_except_level = CRITICAL {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.846496] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] rate_limit_interval = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.846659] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] reboot_timeout = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.846809] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] reclaim_instance_interval = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.846963] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] record = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.847140] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] reimage_timeout_per_gb = 60 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.847307] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] report_interval = 120 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.847467] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] rescue_timeout = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.847626] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] reserved_host_cpus = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.847780] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] reserved_host_disk_mb = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.847934] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] reserved_host_memory_mb = 512 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.848101] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] reserved_huge_pages = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.848265] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] resize_confirm_window = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.848421] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] resize_fs_using_block_device = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.848579] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] resume_guests_state_on_host_boot = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.848745] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.848906] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] rpc_response_timeout = 60 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.849076] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] run_external_periodic_tasks = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.849245] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] running_deleted_instance_action = reap {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.849406] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] running_deleted_instance_poll_interval = 1800 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.849562] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] running_deleted_instance_timeout = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.849720] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] scheduler_instance_sync_interval = 120 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.849882] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] service_down_time = 720 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.850059] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] servicegroup_driver = db {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.850218] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] shell_completion = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.850376] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] shelved_offload_time = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.850533] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] shelved_poll_interval = 3600 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.850693] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] shutdown_timeout = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.850852] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] source_is_ipv6 = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.851014] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ssl_only = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.851264] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.851434] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] sync_power_state_interval = 600 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.851590] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] sync_power_state_pool_size = 1000 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.851755] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] syslog_log_facility = LOG_USER {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.851911] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] tempdir = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.852080] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] timeout_nbd = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.852248] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] transport_url = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.852410] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] update_resources_interval = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.852565] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] use_cow_images = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.852720] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] use_eventlog = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.852878] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] use_journal = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.853042] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] use_json = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.853203] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] use_rootwrap_daemon = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.853359] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] use_stderr = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.853513] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] use_syslog = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.853667] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vcpu_pin_set = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.853830] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vif_plugging_is_fatal = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.853992] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vif_plugging_timeout = 300 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.854169] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] virt_mkfs = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.854331] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] volume_usage_poll_interval = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.854489] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] watch_log_file = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.854654] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] web = /usr/share/spice-html5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 502.854835] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.855007] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.855179] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.855352] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_concurrency.disable_process_locking = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.855895] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.856098] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.856274] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.856454] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.856629] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.856799] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.856978] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api.auth_strategy = keystone {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.857160] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api.compute_link_prefix = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.857339] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.857512] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api.dhcp_domain = novalocal {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.857680] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api.enable_instance_password = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.857845] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api.glance_link_prefix = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.858030] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.858219] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.858380] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api.instance_list_per_project_cells = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.858544] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api.list_records_by_skipping_down_cells = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.858705] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api.local_metadata_per_cell = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.858870] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api.max_limit = 1000 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.859046] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api.metadata_cache_expiration = 15 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.859227] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api.neutron_default_tenant_id = default {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.859401] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api.response_validation = warn {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.859575] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api.use_neutron_default_nets = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.859739] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.859901] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.860076] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.860255] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.860428] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api.vendordata_dynamic_targets = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.860592] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api.vendordata_jsonfile_path = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.860772] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.860964] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.backend = dogpile.cache.memcached {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.861146] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.backend_argument = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.861320] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.config_prefix = cache.oslo {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.861489] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.dead_timeout = 60.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.861652] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.debug_cache_backend = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.861813] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.enable_retry_client = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.861972] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.enable_socket_keepalive = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.862153] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.enabled = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.862318] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.enforce_fips_mode = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.862479] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.expiration_time = 600 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.862640] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.hashclient_retry_attempts = 2 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.862806] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.hashclient_retry_delay = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.862968] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.memcache_dead_retry = 300 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.863139] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.memcache_password = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.863304] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.863464] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.863623] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.memcache_pool_maxsize = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.863783] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.863942] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.memcache_sasl_enabled = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.864127] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.864294] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.memcache_socket_timeout = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.864454] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.memcache_username = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.864617] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.proxies = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.864778] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.redis_db = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.864937] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.redis_password = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.865118] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.redis_sentinel_service_name = mymaster {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.865296] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.865465] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.redis_server = localhost:6379 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.865630] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.redis_socket_timeout = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.865790] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.redis_username = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.865951] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.retry_attempts = 2 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.866152] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.retry_delay = 0.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.866288] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.socket_keepalive_count = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.866451] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.socket_keepalive_idle = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.866611] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.socket_keepalive_interval = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.866766] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.tls_allowed_ciphers = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.866923] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.tls_cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.867091] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.tls_certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.867253] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.tls_enabled = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.867410] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cache.tls_keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.867580] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cinder.auth_section = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.867751] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cinder.auth_type = password {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.867911] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cinder.cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.868095] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cinder.catalog_info = volumev3::publicURL {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.868264] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cinder.certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.868426] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cinder.collect_timing = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.868586] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cinder.cross_az_attach = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.868747] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cinder.debug = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.868905] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cinder.endpoint_template = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.869075] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cinder.http_retries = 3 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.869241] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cinder.insecure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.869401] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cinder.keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.869570] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cinder.os_region_name = RegionOne {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.869735] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cinder.split_loggers = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.869893] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cinder.timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.870075] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.870240] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] compute.cpu_dedicated_set = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.870407] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] compute.cpu_shared_set = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.870574] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] compute.image_type_exclude_list = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.870736] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.870896] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] compute.max_concurrent_disk_ops = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.871066] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] compute.max_disk_devices_to_attach = -1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.871234] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.871404] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.871565] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] compute.resource_provider_association_refresh = 300 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.871724] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.871883] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] compute.shutdown_retry_interval = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.873179] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.873409] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] conductor.workers = 2 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.873604] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] console.allowed_origins = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.877254] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] console.ssl_ciphers = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.877457] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] console.ssl_minimum_version = default {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.879543] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] consoleauth.enforce_session_timeout = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.879543] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] consoleauth.token_ttl = 600 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.879543] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cyborg.cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.879543] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cyborg.certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.879543] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cyborg.collect_timing = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.879543] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cyborg.connect_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.879543] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cyborg.connect_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.879839] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cyborg.endpoint_override = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.879839] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cyborg.insecure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.879839] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cyborg.keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.879839] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cyborg.max_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.879839] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cyborg.min_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.879839] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cyborg.region_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.879839] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cyborg.retriable_status_codes = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.880017] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cyborg.service_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.880017] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cyborg.service_type = accelerator {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.880192] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cyborg.split_loggers = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.880359] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cyborg.status_code_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.880519] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cyborg.status_code_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.880676] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cyborg.timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.880858] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.881032] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] cyborg.version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.881220] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] database.backend = sqlalchemy {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.881394] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] database.connection = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.881561] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] database.connection_debug = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.881731] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] database.connection_parameters = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.881897] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] database.connection_recycle_time = 3600 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.882071] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] database.connection_trace = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.882235] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] database.db_inc_retry_interval = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.882400] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] database.db_max_retries = 20 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.882563] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] database.db_max_retry_interval = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.882726] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] database.db_retry_interval = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.882889] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] database.max_overflow = 50 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.883058] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] database.max_pool_size = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.883224] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] database.max_retries = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.883396] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.883554] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] database.mysql_wsrep_sync_wait = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.883713] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] database.pool_timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.883875] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] database.retry_interval = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.884042] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] database.slave_connection = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.884210] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] database.sqlite_synchronous = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.884376] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] database.use_db_reconnect = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.884551] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api_database.backend = sqlalchemy {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.884720] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api_database.connection = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.884883] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api_database.connection_debug = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.885063] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api_database.connection_parameters = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.885232] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api_database.connection_recycle_time = 3600 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.885398] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api_database.connection_trace = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.885560] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api_database.db_inc_retry_interval = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.885727] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api_database.db_max_retries = 20 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.885882] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api_database.db_max_retry_interval = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.886053] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api_database.db_retry_interval = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.886223] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api_database.max_overflow = 50 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.886382] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api_database.max_pool_size = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.886541] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api_database.max_retries = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.886708] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.886865] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.887032] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api_database.pool_timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.887234] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api_database.retry_interval = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.887356] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api_database.slave_connection = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.887519] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] api_database.sqlite_synchronous = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.887692] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] devices.enabled_mdev_types = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.887868] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.888050] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ephemeral_storage_encryption.default_format = luks {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.888227] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ephemeral_storage_encryption.enabled = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.888385] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.888556] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.api_servers = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.888719] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.888879] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.889053] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.collect_timing = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.889221] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.connect_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.889379] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.connect_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.889541] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.debug = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.889747] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.default_trusted_certificate_ids = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.889862] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.enable_certificate_validation = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.890031] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.enable_rbd_download = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.890195] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.endpoint_override = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.890362] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.insecure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.890525] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.890683] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.max_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.890839] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.min_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.891007] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.num_retries = 3 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.891184] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.rbd_ceph_conf = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.891350] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.rbd_connect_timeout = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.891518] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.rbd_pool = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.891687] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.rbd_user = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.891845] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.region_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.892011] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.retriable_status_codes = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.892178] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.service_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.892348] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.service_type = image {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.892510] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.split_loggers = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.892674] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.status_code_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.892832] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.status_code_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.892990] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.893184] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.893352] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.verify_glance_signatures = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.893512] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] glance.version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.893676] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] guestfs.debug = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.893840] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] mks.enabled = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.894208] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.894405] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] image_cache.manager_interval = 2400 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.894576] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] image_cache.precache_concurrency = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.894749] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] image_cache.remove_unused_base_images = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.894919] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.895100] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.895282] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] image_cache.subdirectory_name = _base {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.895464] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.api_max_retries = 60 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.895635] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.api_retry_interval = 2 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.895794] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.auth_section = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.895955] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.auth_type = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.896127] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.896291] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.896454] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.collect_timing = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.896618] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.conductor_group = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.896776] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.connect_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.896937] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.connect_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.897109] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.endpoint_override = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.897281] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.insecure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.897437] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.897595] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.max_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.897753] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.min_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.897918] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.peer_list = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.898089] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.region_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.898256] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.retriable_status_codes = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.898420] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.serial_console_state_timeout = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.898580] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.service_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.898749] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.service_type = baremetal {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.898908] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.shard = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.899086] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.split_loggers = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.899249] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.status_code_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.899409] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.status_code_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.899567] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.899746] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.899908] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ironic.version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.900099] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.900285] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] key_manager.fixed_key = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.900475] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.900637] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican.barbican_api_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.900797] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican.barbican_endpoint = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.900966] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican.barbican_endpoint_type = public {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.901140] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican.barbican_region_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.901302] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican.cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.901463] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican.certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.901625] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican.collect_timing = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.901791] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican.insecure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.901941] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican.keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.902116] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican.number_of_retries = 60 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.902281] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican.retry_delay = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.902445] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican.send_service_user_token = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.902608] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican.split_loggers = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.902764] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican.timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.902924] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican.verify_ssl = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.903092] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican.verify_ssl_path = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.903260] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican_service_user.auth_section = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.903426] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican_service_user.auth_type = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.903583] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican_service_user.cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.903741] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican_service_user.certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.903903] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican_service_user.collect_timing = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.904074] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican_service_user.insecure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.904237] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican_service_user.keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.904410] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican_service_user.split_loggers = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.904563] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] barbican_service_user.timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.904729] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vault.approle_role_id = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.904891] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vault.approle_secret_id = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.905075] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vault.kv_mountpoint = secret {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.905243] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vault.kv_path = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.905411] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vault.kv_version = 2 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.905570] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vault.namespace = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.905730] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vault.root_token_id = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.905887] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vault.ssl_ca_crt_file = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.906062] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vault.timeout = 60.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.906230] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vault.use_ssl = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.906399] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.906574] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] keystone.auth_section = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.906737] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] keystone.auth_type = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.906894] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] keystone.cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.907065] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] keystone.certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.907232] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] keystone.collect_timing = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.907393] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] keystone.connect_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.907551] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] keystone.connect_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.907709] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] keystone.endpoint_override = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.907870] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] keystone.insecure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.908036] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] keystone.keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.908207] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] keystone.max_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.908356] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] keystone.min_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.908510] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] keystone.region_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.908664] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] keystone.retriable_status_codes = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.908817] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] keystone.service_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.908986] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] keystone.service_type = identity {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.909246] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] keystone.split_loggers = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.909332] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] keystone.status_code_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.909478] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] keystone.status_code_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.909633] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] keystone.timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.909814] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.909975] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] keystone.version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.910189] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.connection_uri = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.910353] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.cpu_mode = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.910519] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.cpu_model_extra_flags = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.910689] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.cpu_models = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.910859] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.cpu_power_governor_high = performance {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.911039] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.cpu_power_governor_low = powersave {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.911209] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.cpu_power_management = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.911383] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.911547] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.device_detach_attempts = 8 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.911710] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.device_detach_timeout = 20 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.911873] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.disk_cachemodes = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.912040] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.disk_prefix = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.912210] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.enabled_perf_events = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.912374] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.file_backed_memory = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.912541] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.gid_maps = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.912697] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.hw_disk_discard = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.912853] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.hw_machine_type = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.913041] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.images_rbd_ceph_conf = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.913212] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.913386] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.913550] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.images_rbd_glance_store_name = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.913720] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.images_rbd_pool = rbd {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.913892] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.images_type = default {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.914061] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.images_volume_group = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.914228] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.inject_key = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.914392] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.inject_partition = -2 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.914552] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.inject_password = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.914714] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.iscsi_iface = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.914877] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.iser_use_multipath = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.915048] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.live_migration_bandwidth = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.915216] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.915381] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.live_migration_downtime = 500 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.915540] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.915703] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.915862] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.live_migration_inbound_addr = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.916054] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.916238] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.live_migration_permit_post_copy = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.916404] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.live_migration_scheme = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.916581] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.live_migration_timeout_action = abort {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.916746] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.live_migration_tunnelled = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.916907] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.live_migration_uri = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.917083] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.live_migration_with_native_tls = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.917246] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.max_queues = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.917412] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.917652] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.917820] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.nfs_mount_options = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.918130] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.918309] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.918475] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.num_iser_scan_tries = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.918637] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.num_memory_encrypted_guests = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.918801] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.918962] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.num_pcie_ports = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.919146] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.num_volume_scan_tries = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.919312] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.pmem_namespaces = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.919472] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.quobyte_client_cfg = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.919754] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.919926] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.rbd_connect_timeout = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.920105] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.920271] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.920433] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.rbd_secret_uuid = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.920593] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.rbd_user = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.920754] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.920928] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.remote_filesystem_transport = ssh {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.921099] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.rescue_image_id = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.921261] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.rescue_kernel_id = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.921420] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.rescue_ramdisk_id = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.921589] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.921747] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.rx_queue_size = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.921913] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.smbfs_mount_options = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.922195] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.922369] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.snapshot_compression = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.922529] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.snapshot_image_format = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.922744] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.922909] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.sparse_logical_volumes = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.923080] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.swtpm_enabled = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.923253] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.swtpm_group = tss {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.923421] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.swtpm_user = tss {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.923586] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.sysinfo_serial = unique {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.923744] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.tb_cache_size = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.923901] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.tx_queue_size = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.924076] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.uid_maps = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.924243] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.use_virtio_for_bridges = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.924414] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.virt_type = kvm {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.924583] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.volume_clear = zero {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.924746] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.volume_clear_size = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.924909] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.volume_use_multipath = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.925076] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.vzstorage_cache_path = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.925247] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.925414] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.vzstorage_mount_group = qemu {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.925579] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.vzstorage_mount_opts = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.925746] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.926024] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.926207] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.vzstorage_mount_user = stack {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.926374] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.926547] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.auth_section = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.926719] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.auth_type = password {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.926880] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.927046] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.927212] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.collect_timing = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.927370] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.connect_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.927527] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.connect_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.927695] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.default_floating_pool = public {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.927855] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.endpoint_override = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.928019] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.extension_sync_interval = 600 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.928195] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.http_retries = 3 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.928347] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.insecure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.928503] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.928663] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.max_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.928831] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.928992] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.min_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.929177] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.ovs_bridge = br-int {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.929346] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.physnets = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.929515] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.region_name = RegionOne {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.929675] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.retriable_status_codes = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.929841] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.service_metadata_proxy = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.930008] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.service_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.930183] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.service_type = network {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.930347] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.split_loggers = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.930505] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.status_code_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.930664] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.status_code_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.930820] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.930998] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.931171] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] neutron.version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.931341] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] notifications.bdms_in_notifications = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.931516] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] notifications.default_level = INFO {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.931686] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] notifications.notification_format = unversioned {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.931848] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] notifications.notify_on_state_change = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.932027] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.932204] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] pci.alias = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.932374] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] pci.device_spec = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.932538] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] pci.report_in_placement = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.932709] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.auth_section = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.932879] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.auth_type = password {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.933053] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.933218] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.933375] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.933538] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.collect_timing = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.933696] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.connect_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.933851] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.connect_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.934014] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.default_domain_id = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.934173] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.default_domain_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.934328] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.domain_id = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.934481] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.domain_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.934637] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.endpoint_override = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.934794] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.insecure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.934949] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.935119] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.max_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.935276] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.min_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.935440] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.password = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.935598] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.project_domain_id = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.935760] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.project_domain_name = Default {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.935927] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.project_id = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.936106] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.project_name = service {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.936276] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.region_name = RegionOne {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.936439] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.retriable_status_codes = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.936599] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.service_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.936767] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.service_type = placement {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.936929] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.split_loggers = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.937101] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.status_code_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.937264] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.status_code_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.937424] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.system_scope = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.937579] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.937768] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.trust_id = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.937891] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.user_domain_id = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.938066] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.user_domain_name = Default {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.938232] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.user_id = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.938404] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.username = nova {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.938583] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.938744] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] placement.version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.938918] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] quota.cores = 20 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.939093] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] quota.count_usage_from_placement = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.939271] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.939445] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] quota.injected_file_content_bytes = 10240 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.939610] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] quota.injected_file_path_length = 255 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.939774] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] quota.injected_files = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.939937] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] quota.instances = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.940108] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] quota.key_pairs = 100 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.940274] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] quota.metadata_items = 128 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.940437] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] quota.ram = 51200 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.940598] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] quota.recheck_quota = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.940766] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] quota.server_group_members = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.940932] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] quota.server_groups = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.941118] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.941284] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.941456] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] scheduler.image_metadata_prefilter = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.941619] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.941782] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] scheduler.max_attempts = 3 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.941943] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] scheduler.max_placement_results = 1000 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.942117] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.942281] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] scheduler.query_placement_for_image_type_support = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.942443] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.942616] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] scheduler.workers = 2 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.942788] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.942965] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.943145] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.943318] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.943483] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.943648] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.943813] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.944009] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.944187] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.host_subset_size = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.944351] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.944509] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.944671] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.944834] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.isolated_hosts = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.944996] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.isolated_images = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.945170] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.945329] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.945491] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.945651] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.pci_in_placement = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.945812] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.945971] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.946143] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.946304] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.946464] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.946621] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.946780] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.track_instance_changes = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.946952] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.947132] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] metrics.required = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.947292] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] metrics.weight_multiplier = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.947454] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.947614] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] metrics.weight_setting = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.947919] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.948103] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] serial_console.enabled = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.948288] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] serial_console.port_range = 10000:20000 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.948461] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.948628] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.948793] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] serial_console.serialproxy_port = 6083 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.948962] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] service_user.auth_section = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.949142] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] service_user.auth_type = password {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.949302] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] service_user.cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.949459] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] service_user.certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.949617] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] service_user.collect_timing = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.949774] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] service_user.insecure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.949928] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] service_user.keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.950103] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] service_user.send_service_user_token = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.950271] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] service_user.split_loggers = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.950440] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] service_user.timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.950607] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] spice.agent_enabled = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.950767] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] spice.enabled = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.951083] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.951279] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.951453] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] spice.html5proxy_port = 6082 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.951616] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] spice.image_compression = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.951776] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] spice.jpeg_compression = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.951935] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] spice.playback_compression = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.952112] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] spice.require_secure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.952284] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] spice.server_listen = 127.0.0.1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.952456] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.952614] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] spice.streaming_mode = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.952773] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] spice.zlib_compression = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.952938] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] upgrade_levels.baseapi = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.953117] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] upgrade_levels.compute = auto {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.953279] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] upgrade_levels.conductor = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.953437] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] upgrade_levels.scheduler = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.953600] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vendordata_dynamic_auth.auth_section = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.953759] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vendordata_dynamic_auth.auth_type = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.953913] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vendordata_dynamic_auth.cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.954076] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vendordata_dynamic_auth.certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.954241] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.954401] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vendordata_dynamic_auth.insecure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.954558] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vendordata_dynamic_auth.keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.954718] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.954871] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vendordata_dynamic_auth.timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.955050] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vmware.api_retry_count = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.955213] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vmware.ca_file = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.955384] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vmware.cache_prefix = devstack-image-cache {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.955549] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vmware.cluster_name = testcl1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.955711] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vmware.connection_pool_size = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.955870] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vmware.console_delay_seconds = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.956047] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vmware.datastore_regex = ^datastore.* {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.956252] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.956422] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vmware.host_password = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.956588] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vmware.host_port = 443 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.956753] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vmware.host_username = administrator@vsphere.local {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.956918] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vmware.insecure = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.957092] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vmware.integration_bridge = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.957255] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vmware.maximum_objects = 100 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.957415] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vmware.pbm_default_policy = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.957576] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vmware.pbm_enabled = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.957733] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vmware.pbm_wsdl_location = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.957899] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.958064] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vmware.serial_port_proxy_uri = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.958226] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vmware.serial_port_service_uri = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.958389] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vmware.task_poll_interval = 0.5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.958553] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vmware.use_linked_clone = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.958722] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vmware.vnc_keymap = en-us {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.958885] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vmware.vnc_port = 5900 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.959056] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vmware.vnc_port_total = 10000 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.959244] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vnc.auth_schemes = ['none'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.959417] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vnc.enabled = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.959705] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.959888] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.960071] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vnc.novncproxy_port = 6080 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.960248] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vnc.server_listen = 127.0.0.1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.960419] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.960581] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vnc.vencrypt_ca_certs = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.960741] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vnc.vencrypt_client_cert = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.960900] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vnc.vencrypt_client_key = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.961089] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.961257] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] workarounds.disable_deep_image_inspection = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.961420] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.961580] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.961738] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.961898] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] workarounds.disable_rootwrap = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.962069] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] workarounds.enable_numa_live_migration = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.962231] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.962394] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.962552] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.962708] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] workarounds.libvirt_disable_apic = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.962865] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.963035] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.963198] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.963358] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.963516] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.963673] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.963829] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.963987] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.964160] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.964326] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.964508] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.964675] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] wsgi.client_socket_timeout = 900 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.964837] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] wsgi.default_pool_size = 1000 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.964998] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] wsgi.keep_alive = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.965179] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] wsgi.max_header_line = 16384 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.965342] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] wsgi.secure_proxy_ssl_header = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.965502] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] wsgi.ssl_ca_file = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.965664] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] wsgi.ssl_cert_file = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.965824] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] wsgi.ssl_key_file = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.965988] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] wsgi.tcp_keepidle = 600 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.966183] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.966355] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] zvm.ca_file = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.966515] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] zvm.cloud_connector_url = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.966789] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.966961] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] zvm.reachable_timeout = 300 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.967155] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_policy.enforce_new_defaults = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.967533] env[61906]: WARNING oslo_config.cfg [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 502.967713] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_policy.enforce_scope = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.967892] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_policy.policy_default_rule = default {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.968086] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.968269] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_policy.policy_file = policy.yaml {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.968449] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.968614] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.968775] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.968933] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.969108] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.969280] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.969453] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.969627] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] profiler.connection_string = messaging:// {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.969793] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] profiler.enabled = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.969963] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] profiler.es_doc_type = notification {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.970138] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] profiler.es_scroll_size = 10000 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.970307] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] profiler.es_scroll_time = 2m {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.970473] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] profiler.filter_error_trace = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.970639] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] profiler.hmac_keys = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.970806] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] profiler.sentinel_service_name = mymaster {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.970972] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] profiler.socket_timeout = 0.1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.971145] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] profiler.trace_requests = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.971308] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] profiler.trace_sqlalchemy = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.971487] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] profiler_jaeger.process_tags = {} {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.971646] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] profiler_jaeger.service_name_prefix = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.971806] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] profiler_otlp.service_name_prefix = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.971967] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] remote_debug.host = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.972137] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] remote_debug.port = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.972318] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.972488] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.972659] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.972820] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.972983] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.973156] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.973321] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.973481] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.973641] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.973808] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.973966] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.974149] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.974316] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.974482] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.974648] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.974813] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.974975] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.975162] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.975328] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.975488] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.975651] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.975812] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.975975] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.976156] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.976319] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.976484] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.976647] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.977607] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.977607] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.977607] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.ssl = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.977607] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.977607] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.977607] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.977817] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.977937] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.ssl_version = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.978115] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.978305] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.978469] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_notifications.retry = -1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.978653] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.978827] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_messaging_notifications.transport_url = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.978998] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_limit.auth_section = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.979174] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_limit.auth_type = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.979864] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_limit.cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.979864] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_limit.certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.979864] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_limit.collect_timing = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.979864] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_limit.connect_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.980120] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_limit.connect_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.980120] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_limit.endpoint_id = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.980256] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_limit.endpoint_override = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.980415] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_limit.insecure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.980571] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_limit.keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.980727] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_limit.max_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.980882] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_limit.min_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984013] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_limit.region_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984013] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_limit.retriable_status_codes = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984013] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_limit.service_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984013] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_limit.service_type = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984013] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_limit.split_loggers = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984013] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_limit.status_code_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984013] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_limit.status_code_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984259] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_limit.timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984259] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_limit.valid_interfaces = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984259] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_limit.version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984259] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_reports.file_event_handler = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984259] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984259] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] oslo_reports.log_dir = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984259] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984436] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984436] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984436] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984436] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984436] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984436] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984436] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vif_plug_ovs_privileged.group = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984602] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984602] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984602] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984868] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] vif_plug_ovs_privileged.user = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.984906] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] os_vif_linux_bridge.flat_interface = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.985082] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.985255] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.985421] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.985587] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.985746] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.985908] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.986079] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.986266] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.986437] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] os_vif_ovs.isolate_vif = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.986602] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.986764] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.986929] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.987109] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] os_vif_ovs.ovsdb_interface = native {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.987270] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] os_vif_ovs.per_port_bridge = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.987437] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] privsep_osbrick.capabilities = [21] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.987595] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] privsep_osbrick.group = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.987748] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] privsep_osbrick.helper_command = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.987910] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.988080] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.988242] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] privsep_osbrick.user = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.988410] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.988566] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] nova_sys_admin.group = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.988721] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] nova_sys_admin.helper_command = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.988880] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.989046] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.989212] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] nova_sys_admin.user = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 502.989947] env[61906]: DEBUG oslo_service.service [None req-3cd9281a-a1b2-4ddc-a40a-a0ef7aad62a0 None None] ******************************************************************************** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 502.989947] env[61906]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 503.492990] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] Getting list of instances from cluster (obj){ [ 503.492990] env[61906]: value = "domain-c8" [ 503.492990] env[61906]: _type = "ClusterComputeResource" [ 503.492990] env[61906]: } {{(pid=61906) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 503.494190] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc2941d-67bc-402e-8e71-9306119879c4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.503400] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] Got total of 0 instances {{(pid=61906) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 503.503925] env[61906]: WARNING nova.virt.vmwareapi.driver [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 503.504418] env[61906]: INFO nova.virt.node [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] Generated node identity a5b57df7-a16e-44f9-9b2d-23c518860263 [ 503.504649] env[61906]: INFO nova.virt.node [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] Wrote node identity a5b57df7-a16e-44f9-9b2d-23c518860263 to /opt/stack/data/n-cpu-1/compute_id [ 504.009560] env[61906]: WARNING nova.compute.manager [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] Compute nodes ['a5b57df7-a16e-44f9-9b2d-23c518860263'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 505.016993] env[61906]: INFO nova.compute.manager [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 506.024075] env[61906]: WARNING nova.compute.manager [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 506.024075] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 506.024075] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 506.024075] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 506.024075] env[61906]: DEBUG nova.compute.resource_tracker [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61906) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 506.024075] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1c5b4d2-30cc-466f-bab7-6a24d41f595e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.032392] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9298c3-b26a-4c61-9390-314c26ff1ceb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.046281] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d20e95-ef2e-4dca-88fa-54142cfdf3f3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.052731] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc71a268-8ea0-4271-87cc-445d2dca378b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.082535] env[61906]: DEBUG nova.compute.resource_tracker [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181447MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=61906) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 506.082858] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 506.083190] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 506.587823] env[61906]: WARNING nova.compute.resource_tracker [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] No compute node record for cpu-1:a5b57df7-a16e-44f9-9b2d-23c518860263: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host a5b57df7-a16e-44f9-9b2d-23c518860263 could not be found. [ 507.090797] env[61906]: INFO nova.compute.resource_tracker [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: a5b57df7-a16e-44f9-9b2d-23c518860263 [ 508.600817] env[61906]: DEBUG nova.compute.resource_tracker [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 508.600817] env[61906]: DEBUG nova.compute.resource_tracker [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 508.753284] env[61906]: INFO nova.scheduler.client.report [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] [req-079b7417-3186-476f-823d-b0131ca2c37d] Created resource provider record via placement API for resource provider with UUID a5b57df7-a16e-44f9-9b2d-23c518860263 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 508.774078] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4b97410-e8ed-4225-b253-7a8036f2ee6c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.780638] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-349c19d3-84b5-40ea-a22e-ae19b65d2c17 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.810609] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d4bfaf5-5a69-42a3-9728-0137f6bfdf33 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.817634] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57ec06f3-616c-4ab0-8b55-64742be539b8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.830498] env[61906]: DEBUG nova.compute.provider_tree [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 509.377212] env[61906]: DEBUG nova.scheduler.client.report [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] Updated inventory for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 509.377212] env[61906]: DEBUG nova.compute.provider_tree [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] Updating resource provider a5b57df7-a16e-44f9-9b2d-23c518860263 generation from 0 to 1 during operation: update_inventory {{(pid=61906) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 509.377212] env[61906]: DEBUG nova.compute.provider_tree [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 509.430726] env[61906]: DEBUG nova.compute.provider_tree [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] Updating resource provider a5b57df7-a16e-44f9-9b2d-23c518860263 generation from 1 to 2 during operation: update_traits {{(pid=61906) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 509.935736] env[61906]: DEBUG nova.compute.resource_tracker [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 509.935736] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.850s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 509.935736] env[61906]: DEBUG nova.service [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] Creating RPC server for service compute {{(pid=61906) start /opt/stack/nova/nova/service.py:186}} [ 509.947675] env[61906]: DEBUG nova.service [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] Join ServiceGroup membership for this service compute {{(pid=61906) start /opt/stack/nova/nova/service.py:203}} [ 509.948106] env[61906]: DEBUG nova.servicegroup.drivers.db [None req-ac1fc563-307b-45eb-8c64-42bd020f0a3f None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=61906) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 547.812125] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Acquiring lock "47f866b0-5752-4439-9dca-de44934955a6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.812525] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Lock "47f866b0-5752-4439-9dca-de44934955a6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.316181] env[61906]: DEBUG nova.compute.manager [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 548.869453] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.870473] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.872472] env[61906]: INFO nova.compute.claims [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 549.612473] env[61906]: DEBUG oslo_concurrency.lockutils [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Acquiring lock "402707da-f091-40fb-b5c7-3a2b1429ecce" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.612727] env[61906]: DEBUG oslo_concurrency.lockutils [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Lock "402707da-f091-40fb-b5c7-3a2b1429ecce" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.614574] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "7044fd5f-a202-45a4-a4fb-73d20ef87cb5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.614797] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "7044fd5f-a202-45a4-a4fb-73d20ef87cb5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.982237] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30dd6433-eb3d-47a6-a34d-8ca42a8f5ec9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.992292] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3764e74-d694-4159-8f2a-1f00d6b01afc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.032985] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb75deed-a64b-4e79-b781-7178226d3e9b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.043077] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b95d3cae-2976-40cc-930a-8441b9d48f1e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.058692] env[61906]: DEBUG nova.compute.provider_tree [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 550.119681] env[61906]: DEBUG nova.compute.manager [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 550.133395] env[61906]: DEBUG nova.compute.manager [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 550.271060] env[61906]: DEBUG oslo_concurrency.lockutils [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Acquiring lock "739af291-794c-491d-b579-2c50c5a14785" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.272098] env[61906]: DEBUG oslo_concurrency.lockutils [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Lock "739af291-794c-491d-b579-2c50c5a14785" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.562583] env[61906]: DEBUG nova.scheduler.client.report [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 550.651470] env[61906]: DEBUG oslo_concurrency.lockutils [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.662270] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.778167] env[61906]: DEBUG nova.compute.manager [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 550.884841] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Acquiring lock "1ac58495-75e5-4bce-8cb9-ecf18e0e436b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.885112] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Lock "1ac58495-75e5-4bce-8cb9-ecf18e0e436b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.951987] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._sync_power_states {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 551.069934] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.199s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.070938] env[61906]: DEBUG nova.compute.manager [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 551.078351] env[61906]: DEBUG oslo_concurrency.lockutils [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.427s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.081388] env[61906]: INFO nova.compute.claims [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 551.303815] env[61906]: DEBUG oslo_concurrency.lockutils [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.387798] env[61906]: DEBUG nova.compute.manager [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 551.454513] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Getting list of instances from cluster (obj){ [ 551.454513] env[61906]: value = "domain-c8" [ 551.454513] env[61906]: _type = "ClusterComputeResource" [ 551.454513] env[61906]: } {{(pid=61906) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 551.455634] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-386f526c-a332-456a-8328-7ba8c9055857 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.466500] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Got total of 0 instances {{(pid=61906) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 551.466892] env[61906]: WARNING nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] While synchronizing instance power states, found 1 instances in the database and 0 instances on the hypervisor. [ 551.467556] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Triggering sync for uuid 47f866b0-5752-4439-9dca-de44934955a6 {{(pid=61906) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 551.467556] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "47f866b0-5752-4439-9dca-de44934955a6" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.467797] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 551.468376] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Getting list of instances from cluster (obj){ [ 551.468376] env[61906]: value = "domain-c8" [ 551.468376] env[61906]: _type = "ClusterComputeResource" [ 551.468376] env[61906]: } {{(pid=61906) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 551.470149] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9743a230-b5a3-48c9-8bbc-f7f02bed43a9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.480018] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Got total of 0 instances {{(pid=61906) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 551.587033] env[61906]: DEBUG nova.compute.utils [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 551.590818] env[61906]: DEBUG nova.compute.manager [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 551.591225] env[61906]: DEBUG nova.network.neutron [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 551.933231] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.097333] env[61906]: DEBUG nova.compute.manager [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 552.217482] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9150bda5-39cd-4065-90ac-a40e7c6055b2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.226620] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8211af92-c218-419e-acc7-046fd1633df4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.261878] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0267ec3-d352-4ac6-b7b4-dc89d632d383 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.271020] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-452c6805-6f96-45bf-a9b5-e086db2902c7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.284481] env[61906]: DEBUG nova.compute.provider_tree [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 552.402677] env[61906]: DEBUG nova.policy [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c27dd79c49af4940930afc0c319c81c3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '235fbc5b4af543718a21ccfe8db1141e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 552.618492] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Acquiring lock "bf480958-27a3-4450-b09c-cba96375712b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.619021] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Lock "bf480958-27a3-4450-b09c-cba96375712b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.787923] env[61906]: DEBUG nova.scheduler.client.report [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 553.120709] env[61906]: DEBUG nova.compute.manager [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 553.125690] env[61906]: DEBUG nova.compute.manager [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 553.167636] env[61906]: DEBUG nova.virt.hardware [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 553.167854] env[61906]: DEBUG nova.virt.hardware [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 553.174286] env[61906]: DEBUG nova.virt.hardware [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 553.174286] env[61906]: DEBUG nova.virt.hardware [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 553.174286] env[61906]: DEBUG nova.virt.hardware [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 553.174286] env[61906]: DEBUG nova.virt.hardware [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 553.174286] env[61906]: DEBUG nova.virt.hardware [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 553.174561] env[61906]: DEBUG nova.virt.hardware [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 553.174561] env[61906]: DEBUG nova.virt.hardware [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 553.174561] env[61906]: DEBUG nova.virt.hardware [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 553.174561] env[61906]: DEBUG nova.virt.hardware [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 553.174561] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95dcc0bc-41f4-47e6-b1d5-815610235d4a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.185907] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb914e98-c837-401e-b914-9341786bfdfe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.206432] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c242752b-0465-48d0-af94-fb06192bd47e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.293499] env[61906]: DEBUG oslo_concurrency.lockutils [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.215s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.294280] env[61906]: DEBUG nova.compute.manager [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 553.298391] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.636s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.300173] env[61906]: INFO nova.compute.claims [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 553.399492] env[61906]: DEBUG nova.network.neutron [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Successfully created port: 8a65f04d-2ad4-4d14-a8f2-0f2200525b06 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 553.660657] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.809345] env[61906]: DEBUG nova.compute.utils [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 553.814294] env[61906]: DEBUG nova.compute.manager [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 553.815721] env[61906]: DEBUG nova.network.neutron [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 553.924932] env[61906]: DEBUG nova.policy [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '77a3b67945e541ba8b9bb4d27f30fc0f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '12ed4085a06048a58ebdbb84a7bb3d4c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 554.321294] env[61906]: DEBUG nova.compute.manager [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 554.478922] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e499d7a7-2c1a-4752-915b-1703e3ed237f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.491369] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4956312e-62e1-4c65-a7b0-c62c812ebf3b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.534848] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a929ef7-33eb-404d-8750-34be731c3fdf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.546019] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26cb4ad4-26a6-425e-bb88-2cd040eab705 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.559878] env[61906]: DEBUG nova.compute.provider_tree [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 555.063476] env[61906]: DEBUG nova.scheduler.client.report [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 555.333277] env[61906]: DEBUG nova.compute.manager [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 555.375664] env[61906]: DEBUG nova.virt.hardware [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 555.375664] env[61906]: DEBUG nova.virt.hardware [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 555.375664] env[61906]: DEBUG nova.virt.hardware [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 555.375877] env[61906]: DEBUG nova.virt.hardware [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 555.375877] env[61906]: DEBUG nova.virt.hardware [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 555.375877] env[61906]: DEBUG nova.virt.hardware [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 555.375877] env[61906]: DEBUG nova.virt.hardware [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 555.375877] env[61906]: DEBUG nova.virt.hardware [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 555.376042] env[61906]: DEBUG nova.virt.hardware [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 555.376042] env[61906]: DEBUG nova.virt.hardware [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 555.376042] env[61906]: DEBUG nova.virt.hardware [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 555.377024] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ce6a4d3-8ce7-49dc-ab37-ca8ab3ff1f78 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.391496] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39d77c30-e6fd-4675-ada9-bfd3fa3a1907 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.550807] env[61906]: DEBUG nova.network.neutron [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Successfully created port: fee6c75a-e766-4956-a3e2-1149ebf95e92 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 555.569163] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.271s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.569679] env[61906]: DEBUG nova.compute.manager [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 555.572268] env[61906]: DEBUG oslo_concurrency.lockutils [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.269s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.574227] env[61906]: INFO nova.compute.claims [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 556.082810] env[61906]: DEBUG nova.compute.utils [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 556.089898] env[61906]: DEBUG nova.compute.manager [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 556.090338] env[61906]: DEBUG nova.network.neutron [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 556.260129] env[61906]: DEBUG nova.policy [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e89d45df95134c709b569d06744dea5e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1b33dea0794a48f78b9f519cb269a8c2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 556.593459] env[61906]: DEBUG nova.compute.manager [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 556.722272] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8cf8a06-0537-4689-8b6a-e982a046e1fc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.729520] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de314ad8-f49f-4afd-bb62-979483037948 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.769732] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3997862-b14f-408b-ba39-7e1320db13ec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.775941] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5809556-91f8-42b5-909e-1004dfa70628 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.791967] env[61906]: DEBUG nova.compute.provider_tree [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 557.295218] env[61906]: DEBUG nova.scheduler.client.report [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 557.610283] env[61906]: DEBUG nova.compute.manager [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 557.650538] env[61906]: DEBUG nova.virt.hardware [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 557.650737] env[61906]: DEBUG nova.virt.hardware [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 557.650895] env[61906]: DEBUG nova.virt.hardware [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 557.652346] env[61906]: DEBUG nova.virt.hardware [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 557.653199] env[61906]: DEBUG nova.virt.hardware [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 557.653199] env[61906]: DEBUG nova.virt.hardware [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 557.653199] env[61906]: DEBUG nova.virt.hardware [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 557.653199] env[61906]: DEBUG nova.virt.hardware [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 557.653386] env[61906]: DEBUG nova.virt.hardware [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 557.653386] env[61906]: DEBUG nova.virt.hardware [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 557.653541] env[61906]: DEBUG nova.virt.hardware [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 557.654438] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6329a2b-5dfa-4563-be3a-67ce364277a2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.668686] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7a0aa7c-8856-41fd-81a4-0abeb80322ae {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.804970] env[61906]: DEBUG oslo_concurrency.lockutils [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.233s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.805519] env[61906]: DEBUG nova.compute.manager [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 557.809741] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.877s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.811187] env[61906]: INFO nova.compute.claims [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 558.079064] env[61906]: ERROR nova.compute.manager [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8a65f04d-2ad4-4d14-a8f2-0f2200525b06, please check neutron logs for more information. [ 558.079064] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 558.079064] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 558.079064] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 558.079064] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 558.079064] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 558.079064] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 558.079064] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 558.079064] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.079064] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 558.079064] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.079064] env[61906]: ERROR nova.compute.manager raise self.value [ 558.079064] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 558.079064] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 558.079064] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.079064] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 558.079609] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.079609] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 558.079609] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8a65f04d-2ad4-4d14-a8f2-0f2200525b06, please check neutron logs for more information. [ 558.079609] env[61906]: ERROR nova.compute.manager [ 558.079609] env[61906]: Traceback (most recent call last): [ 558.079609] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 558.079609] env[61906]: listener.cb(fileno) [ 558.079609] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 558.079609] env[61906]: result = function(*args, **kwargs) [ 558.079609] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 558.079609] env[61906]: return func(*args, **kwargs) [ 558.079609] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 558.079609] env[61906]: raise e [ 558.079609] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 558.079609] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 558.079609] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 558.079609] env[61906]: created_port_ids = self._update_ports_for_instance( [ 558.079609] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 558.079609] env[61906]: with excutils.save_and_reraise_exception(): [ 558.079609] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.079609] env[61906]: self.force_reraise() [ 558.079609] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.079609] env[61906]: raise self.value [ 558.079609] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 558.079609] env[61906]: updated_port = self._update_port( [ 558.079609] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.079609] env[61906]: _ensure_no_port_binding_failure(port) [ 558.079609] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.079609] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 558.081567] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 8a65f04d-2ad4-4d14-a8f2-0f2200525b06, please check neutron logs for more information. [ 558.081567] env[61906]: Removing descriptor: 15 [ 558.084356] env[61906]: ERROR nova.compute.manager [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8a65f04d-2ad4-4d14-a8f2-0f2200525b06, please check neutron logs for more information. [ 558.084356] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] Traceback (most recent call last): [ 558.084356] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 558.084356] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] yield resources [ 558.084356] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 558.084356] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] self.driver.spawn(context, instance, image_meta, [ 558.084356] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 558.084356] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 558.084356] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 558.084356] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] vm_ref = self.build_virtual_machine(instance, [ 558.084356] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 558.084924] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] vif_infos = vmwarevif.get_vif_info(self._session, [ 558.084924] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 558.084924] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] for vif in network_info: [ 558.084924] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 558.084924] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] return self._sync_wrapper(fn, *args, **kwargs) [ 558.084924] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 558.084924] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] self.wait() [ 558.084924] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 558.084924] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] self[:] = self._gt.wait() [ 558.084924] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 558.084924] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] return self._exit_event.wait() [ 558.084924] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 558.084924] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] result = hub.switch() [ 558.085424] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 558.085424] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] return self.greenlet.switch() [ 558.085424] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 558.085424] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] result = function(*args, **kwargs) [ 558.085424] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 558.085424] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] return func(*args, **kwargs) [ 558.085424] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 558.085424] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] raise e [ 558.085424] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 558.085424] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] nwinfo = self.network_api.allocate_for_instance( [ 558.085424] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 558.085424] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] created_port_ids = self._update_ports_for_instance( [ 558.085424] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 558.086592] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] with excutils.save_and_reraise_exception(): [ 558.086592] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.086592] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] self.force_reraise() [ 558.086592] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.086592] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] raise self.value [ 558.086592] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 558.086592] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] updated_port = self._update_port( [ 558.086592] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.086592] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] _ensure_no_port_binding_failure(port) [ 558.086592] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.086592] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] raise exception.PortBindingFailed(port_id=port['id']) [ 558.086592] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] nova.exception.PortBindingFailed: Binding failed for port 8a65f04d-2ad4-4d14-a8f2-0f2200525b06, please check neutron logs for more information. [ 558.086592] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] [ 558.087291] env[61906]: INFO nova.compute.manager [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Terminating instance [ 558.087409] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Acquiring lock "refresh_cache-47f866b0-5752-4439-9dca-de44934955a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 558.087879] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Acquired lock "refresh_cache-47f866b0-5752-4439-9dca-de44934955a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 558.089684] env[61906]: DEBUG nova.network.neutron [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 558.320375] env[61906]: DEBUG nova.compute.utils [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 558.329403] env[61906]: DEBUG nova.compute.manager [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 558.329626] env[61906]: DEBUG nova.network.neutron [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 558.388481] env[61906]: DEBUG nova.network.neutron [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Successfully created port: d49dae9f-3dfd-40e8-8835-ec36ed215ed8 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 558.401456] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 558.401803] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 558.401996] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Starting heal instance info cache {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 558.402238] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Rebuilding the list of instances to heal {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 558.654475] env[61906]: DEBUG nova.policy [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f84222ffa2a343778a2644a66a2ea5c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c179b15b7134636afdc37f35c3e33cb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 558.681199] env[61906]: DEBUG nova.network.neutron [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.831744] env[61906]: DEBUG nova.compute.manager [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 558.841975] env[61906]: DEBUG nova.compute.manager [req-e84d2f36-648c-4690-9065-e0ca0fff4a79 req-6a7f6f7e-09ac-4519-b39b-6e51934befa8 service nova] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Received event network-changed-8a65f04d-2ad4-4d14-a8f2-0f2200525b06 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 558.841975] env[61906]: DEBUG nova.compute.manager [req-e84d2f36-648c-4690-9065-e0ca0fff4a79 req-6a7f6f7e-09ac-4519-b39b-6e51934befa8 service nova] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Refreshing instance network info cache due to event network-changed-8a65f04d-2ad4-4d14-a8f2-0f2200525b06. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 558.841975] env[61906]: DEBUG oslo_concurrency.lockutils [req-e84d2f36-648c-4690-9065-e0ca0fff4a79 req-6a7f6f7e-09ac-4519-b39b-6e51934befa8 service nova] Acquiring lock "refresh_cache-47f866b0-5752-4439-9dca-de44934955a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 558.868082] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Acquiring lock "75ea63d7-6fc7-4a9f-b3df-05c6d12fe913" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.868225] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Lock "75ea63d7-6fc7-4a9f-b3df-05c6d12fe913" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.906086] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 558.906251] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 558.906510] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 558.906510] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 739af291-794c-491d-b579-2c50c5a14785] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 558.906646] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 558.906783] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Didn't find any instances for network info cache update. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 558.907327] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 558.907501] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 558.908220] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 558.908705] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 558.908915] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 558.909404] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 558.909404] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61906) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 558.909496] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 559.008941] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e64eda1-ff03-4e1c-958c-6ba485fb0509 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.015084] env[61906]: DEBUG nova.network.neutron [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.023925] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6e23692-0f94-454e-a841-5f9fecbb9c4e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.064074] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d17b2665-9ded-4cb3-be95-f9af62d15fc5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.072702] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89b5c5ab-18b9-4c17-aade-566103b9e61c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.087080] env[61906]: DEBUG nova.compute.provider_tree [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 559.377472] env[61906]: DEBUG nova.compute.manager [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 559.412926] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.521065] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Releasing lock "refresh_cache-47f866b0-5752-4439-9dca-de44934955a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.521065] env[61906]: DEBUG nova.compute.manager [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 559.521065] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 559.521065] env[61906]: DEBUG oslo_concurrency.lockutils [req-e84d2f36-648c-4690-9065-e0ca0fff4a79 req-6a7f6f7e-09ac-4519-b39b-6e51934befa8 service nova] Acquired lock "refresh_cache-47f866b0-5752-4439-9dca-de44934955a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.521065] env[61906]: DEBUG nova.network.neutron [req-e84d2f36-648c-4690-9065-e0ca0fff4a79 req-6a7f6f7e-09ac-4519-b39b-6e51934befa8 service nova] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Refreshing network info cache for port 8a65f04d-2ad4-4d14-a8f2-0f2200525b06 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 559.521918] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ed82264a-e776-4d88-9bf5-7402ce505c63 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.536529] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-759e3a53-7b42-4f38-ba4a-86c7e72aefa0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.564995] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 47f866b0-5752-4439-9dca-de44934955a6 could not be found. [ 559.565316] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 559.565877] env[61906]: INFO nova.compute.manager [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Took 0.05 seconds to destroy the instance on the hypervisor. [ 559.566147] env[61906]: DEBUG oslo.service.loopingcall [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 559.566390] env[61906]: DEBUG nova.compute.manager [-] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 559.566553] env[61906]: DEBUG nova.network.neutron [-] [instance: 47f866b0-5752-4439-9dca-de44934955a6] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 559.590875] env[61906]: DEBUG nova.scheduler.client.report [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 559.619438] env[61906]: DEBUG nova.network.neutron [-] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 559.851663] env[61906]: DEBUG nova.compute.manager [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 559.896481] env[61906]: DEBUG nova.virt.hardware [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 559.896481] env[61906]: DEBUG nova.virt.hardware [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 559.896481] env[61906]: DEBUG nova.virt.hardware [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 559.899025] env[61906]: DEBUG nova.virt.hardware [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 559.899025] env[61906]: DEBUG nova.virt.hardware [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 559.899025] env[61906]: DEBUG nova.virt.hardware [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 559.899025] env[61906]: DEBUG nova.virt.hardware [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 559.899025] env[61906]: DEBUG nova.virt.hardware [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 559.899424] env[61906]: DEBUG nova.virt.hardware [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 559.899424] env[61906]: DEBUG nova.virt.hardware [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 559.899485] env[61906]: DEBUG nova.virt.hardware [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 559.901600] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd6aeeb6-1e17-4d0b-9f8b-17e0b3ceb54f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.912846] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d738116-7067-45df-92f0-b168c42faf62 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.936415] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.955790] env[61906]: DEBUG nova.network.neutron [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Successfully created port: 9738c91a-842b-4cad-9068-a96785f62fd1 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 560.050588] env[61906]: ERROR nova.compute.manager [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fee6c75a-e766-4956-a3e2-1149ebf95e92, please check neutron logs for more information. [ 560.050588] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 560.050588] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 560.050588] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 560.050588] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 560.050588] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 560.050588] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 560.050588] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 560.050588] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.050588] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 560.050588] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.050588] env[61906]: ERROR nova.compute.manager raise self.value [ 560.050588] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 560.050588] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 560.050588] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.050588] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 560.051133] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.051133] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 560.051133] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fee6c75a-e766-4956-a3e2-1149ebf95e92, please check neutron logs for more information. [ 560.051133] env[61906]: ERROR nova.compute.manager [ 560.051133] env[61906]: Traceback (most recent call last): [ 560.051133] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 560.051133] env[61906]: listener.cb(fileno) [ 560.051133] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 560.051133] env[61906]: result = function(*args, **kwargs) [ 560.051133] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 560.051133] env[61906]: return func(*args, **kwargs) [ 560.051133] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 560.051133] env[61906]: raise e [ 560.051133] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 560.051133] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 560.051133] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 560.051133] env[61906]: created_port_ids = self._update_ports_for_instance( [ 560.051133] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 560.051133] env[61906]: with excutils.save_and_reraise_exception(): [ 560.051133] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.051133] env[61906]: self.force_reraise() [ 560.051133] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.051133] env[61906]: raise self.value [ 560.051133] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 560.051133] env[61906]: updated_port = self._update_port( [ 560.051133] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.051133] env[61906]: _ensure_no_port_binding_failure(port) [ 560.051133] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.051133] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 560.051875] env[61906]: nova.exception.PortBindingFailed: Binding failed for port fee6c75a-e766-4956-a3e2-1149ebf95e92, please check neutron logs for more information. [ 560.051875] env[61906]: Removing descriptor: 16 [ 560.051875] env[61906]: ERROR nova.compute.manager [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fee6c75a-e766-4956-a3e2-1149ebf95e92, please check neutron logs for more information. [ 560.051875] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Traceback (most recent call last): [ 560.051875] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 560.051875] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] yield resources [ 560.051875] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 560.051875] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] self.driver.spawn(context, instance, image_meta, [ 560.051875] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 560.051875] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] self._vmops.spawn(context, instance, image_meta, injected_files, [ 560.051875] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 560.051875] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] vm_ref = self.build_virtual_machine(instance, [ 560.052248] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 560.052248] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] vif_infos = vmwarevif.get_vif_info(self._session, [ 560.052248] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 560.052248] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] for vif in network_info: [ 560.052248] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 560.052248] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] return self._sync_wrapper(fn, *args, **kwargs) [ 560.052248] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 560.052248] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] self.wait() [ 560.052248] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 560.052248] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] self[:] = self._gt.wait() [ 560.052248] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 560.052248] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] return self._exit_event.wait() [ 560.052248] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 560.052607] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] result = hub.switch() [ 560.052607] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 560.052607] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] return self.greenlet.switch() [ 560.052607] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 560.052607] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] result = function(*args, **kwargs) [ 560.052607] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 560.052607] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] return func(*args, **kwargs) [ 560.052607] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 560.052607] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] raise e [ 560.052607] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 560.052607] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] nwinfo = self.network_api.allocate_for_instance( [ 560.052607] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 560.052607] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] created_port_ids = self._update_ports_for_instance( [ 560.052988] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 560.052988] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] with excutils.save_and_reraise_exception(): [ 560.052988] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.052988] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] self.force_reraise() [ 560.052988] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.052988] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] raise self.value [ 560.052988] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 560.052988] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] updated_port = self._update_port( [ 560.052988] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.052988] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] _ensure_no_port_binding_failure(port) [ 560.052988] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.052988] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] raise exception.PortBindingFailed(port_id=port['id']) [ 560.053309] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] nova.exception.PortBindingFailed: Binding failed for port fee6c75a-e766-4956-a3e2-1149ebf95e92, please check neutron logs for more information. [ 560.053309] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] [ 560.053309] env[61906]: INFO nova.compute.manager [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Terminating instance [ 560.055430] env[61906]: DEBUG oslo_concurrency.lockutils [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Acquiring lock "refresh_cache-402707da-f091-40fb-b5c7-3a2b1429ecce" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.057202] env[61906]: DEBUG oslo_concurrency.lockutils [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Acquired lock "refresh_cache-402707da-f091-40fb-b5c7-3a2b1429ecce" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.057202] env[61906]: DEBUG nova.network.neutron [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 560.096195] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.286s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.096705] env[61906]: DEBUG nova.compute.manager [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 560.099428] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.439s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.101284] env[61906]: INFO nova.compute.claims [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 560.125530] env[61906]: DEBUG nova.network.neutron [-] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.210459] env[61906]: DEBUG nova.network.neutron [req-e84d2f36-648c-4690-9065-e0ca0fff4a79 req-6a7f6f7e-09ac-4519-b39b-6e51934befa8 service nova] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 560.468113] env[61906]: DEBUG nova.network.neutron [req-e84d2f36-648c-4690-9065-e0ca0fff4a79 req-6a7f6f7e-09ac-4519-b39b-6e51934befa8 service nova] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.607017] env[61906]: DEBUG nova.compute.utils [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 560.614183] env[61906]: DEBUG nova.compute.manager [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 560.614183] env[61906]: DEBUG nova.network.neutron [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 560.629722] env[61906]: DEBUG nova.network.neutron [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 560.632225] env[61906]: INFO nova.compute.manager [-] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Took 1.07 seconds to deallocate network for instance. [ 560.635051] env[61906]: DEBUG nova.compute.claims [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 560.635051] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.871695] env[61906]: DEBUG nova.policy [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'efd5701316d44d01afd294fc70e0e332', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'de0603a051d2447aadbfbe91a343fda8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 560.969802] env[61906]: DEBUG oslo_concurrency.lockutils [req-e84d2f36-648c-4690-9065-e0ca0fff4a79 req-6a7f6f7e-09ac-4519-b39b-6e51934befa8 service nova] Releasing lock "refresh_cache-47f866b0-5752-4439-9dca-de44934955a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.097346] env[61906]: DEBUG nova.network.neutron [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.116784] env[61906]: DEBUG nova.compute.manager [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 561.374607] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ca2cef5-0bda-4b9e-b047-8c6b202d12ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.388116] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3be107f2-133e-40e3-8135-10c525d6486b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.422800] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44eac3c7-ea04-4544-bd7a-810eec7ec72e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.430939] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14024dd7-db59-4b10-95a3-4b52dc1efe39 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.445046] env[61906]: DEBUG nova.compute.provider_tree [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 561.602683] env[61906]: DEBUG oslo_concurrency.lockutils [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Releasing lock "refresh_cache-402707da-f091-40fb-b5c7-3a2b1429ecce" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.602683] env[61906]: DEBUG nova.compute.manager [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 561.603014] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 561.603318] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1e3951ec-fac4-4e3b-a25d-ec544561a2ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.606539] env[61906]: DEBUG nova.network.neutron [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Successfully created port: 23b80c78-a492-49a8-a9e9-2d6e03a611fa {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 561.616946] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d9021a-a8cf-4913-b639-c4bc443dff96 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.649851] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 402707da-f091-40fb-b5c7-3a2b1429ecce could not be found. [ 561.649851] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 561.649851] env[61906]: INFO nova.compute.manager [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Took 0.05 seconds to destroy the instance on the hypervisor. [ 561.649851] env[61906]: DEBUG oslo.service.loopingcall [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 561.649851] env[61906]: DEBUG nova.compute.manager [-] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 561.649851] env[61906]: DEBUG nova.network.neutron [-] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 561.722487] env[61906]: DEBUG nova.network.neutron [-] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.779259] env[61906]: DEBUG nova.compute.manager [req-a32a4ebb-8bd8-473b-9e5c-94e248941153 req-cb0231ca-2dc6-4aeb-b6b2-09f4406908fd service nova] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Received event network-changed-fee6c75a-e766-4956-a3e2-1149ebf95e92 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 561.779259] env[61906]: DEBUG nova.compute.manager [req-a32a4ebb-8bd8-473b-9e5c-94e248941153 req-cb0231ca-2dc6-4aeb-b6b2-09f4406908fd service nova] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Refreshing instance network info cache due to event network-changed-fee6c75a-e766-4956-a3e2-1149ebf95e92. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 561.779259] env[61906]: DEBUG oslo_concurrency.lockutils [req-a32a4ebb-8bd8-473b-9e5c-94e248941153 req-cb0231ca-2dc6-4aeb-b6b2-09f4406908fd service nova] Acquiring lock "refresh_cache-402707da-f091-40fb-b5c7-3a2b1429ecce" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.779259] env[61906]: DEBUG oslo_concurrency.lockutils [req-a32a4ebb-8bd8-473b-9e5c-94e248941153 req-cb0231ca-2dc6-4aeb-b6b2-09f4406908fd service nova] Acquired lock "refresh_cache-402707da-f091-40fb-b5c7-3a2b1429ecce" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.779259] env[61906]: DEBUG nova.network.neutron [req-a32a4ebb-8bd8-473b-9e5c-94e248941153 req-cb0231ca-2dc6-4aeb-b6b2-09f4406908fd service nova] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Refreshing network info cache for port fee6c75a-e766-4956-a3e2-1149ebf95e92 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 561.845959] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Acquiring lock "75124494-73f0-4188-817b-1c7b4a85c8de" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.846140] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Lock "75124494-73f0-4188-817b-1c7b4a85c8de" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.949265] env[61906]: DEBUG nova.scheduler.client.report [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 562.139441] env[61906]: DEBUG nova.compute.manager [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 562.176072] env[61906]: DEBUG nova.virt.hardware [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 562.176336] env[61906]: DEBUG nova.virt.hardware [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 562.176590] env[61906]: DEBUG nova.virt.hardware [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 562.180093] env[61906]: DEBUG nova.virt.hardware [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 562.180285] env[61906]: DEBUG nova.virt.hardware [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 562.180443] env[61906]: DEBUG nova.virt.hardware [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 562.180656] env[61906]: DEBUG nova.virt.hardware [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 562.180939] env[61906]: DEBUG nova.virt.hardware [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 562.181288] env[61906]: DEBUG nova.virt.hardware [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 562.182388] env[61906]: DEBUG nova.virt.hardware [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 562.183380] env[61906]: DEBUG nova.virt.hardware [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 562.184608] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-342f5fe7-45e8-42f4-ad7d-ae1985577bc7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.200341] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b3707a1-0ee9-46d9-9e5d-b7b2ce8430fa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.225374] env[61906]: DEBUG nova.network.neutron [-] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.241079] env[61906]: DEBUG nova.compute.manager [req-2a6601ed-4054-4ab4-8afa-f437b353f82c req-3e6d1e6d-35df-464c-855e-9d47cb101a60 service nova] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Received event network-vif-deleted-8a65f04d-2ad4-4d14-a8f2-0f2200525b06 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 562.337924] env[61906]: DEBUG nova.network.neutron [req-a32a4ebb-8bd8-473b-9e5c-94e248941153 req-cb0231ca-2dc6-4aeb-b6b2-09f4406908fd service nova] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 562.349391] env[61906]: DEBUG nova.compute.manager [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 562.456610] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.357s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.457745] env[61906]: DEBUG nova.compute.manager [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 562.461976] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 3.049s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.464049] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.464049] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61906) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 562.464049] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.527s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.469913] env[61906]: INFO nova.compute.claims [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 562.476972] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73c5e74b-2e3f-4310-a2f5-f911a515c020 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.499791] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa59c85-8a5b-45bf-83dc-b22ec0241690 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.516660] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e9a1fb6-0cb5-489e-bc45-559e99d13316 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.527468] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77c7ec84-c797-4cec-80c7-ac5ec72d86ea {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.584862] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181442MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=61906) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 562.584862] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.632749] env[61906]: DEBUG nova.network.neutron [req-a32a4ebb-8bd8-473b-9e5c-94e248941153 req-cb0231ca-2dc6-4aeb-b6b2-09f4406908fd service nova] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.727684] env[61906]: INFO nova.compute.manager [-] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Took 1.08 seconds to deallocate network for instance. [ 562.733992] env[61906]: DEBUG nova.compute.claims [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 562.733992] env[61906]: DEBUG oslo_concurrency.lockutils [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.880375] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.978036] env[61906]: DEBUG nova.compute.utils [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 562.979110] env[61906]: DEBUG nova.compute.manager [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 562.979428] env[61906]: DEBUG nova.network.neutron [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 563.136329] env[61906]: DEBUG nova.policy [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0d3381612c1b406e87b72bd623eeb3d3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e6eca5877fdd43bbb36a6ec9f826acc8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 563.138665] env[61906]: DEBUG oslo_concurrency.lockutils [req-a32a4ebb-8bd8-473b-9e5c-94e248941153 req-cb0231ca-2dc6-4aeb-b6b2-09f4406908fd service nova] Releasing lock "refresh_cache-402707da-f091-40fb-b5c7-3a2b1429ecce" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.315872] env[61906]: ERROR nova.compute.manager [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d49dae9f-3dfd-40e8-8835-ec36ed215ed8, please check neutron logs for more information. [ 563.315872] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 563.315872] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.315872] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 563.315872] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.315872] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 563.315872] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.315872] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 563.315872] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.315872] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 563.315872] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.315872] env[61906]: ERROR nova.compute.manager raise self.value [ 563.315872] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.315872] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 563.315872] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.315872] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 563.316398] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.316398] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 563.316398] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d49dae9f-3dfd-40e8-8835-ec36ed215ed8, please check neutron logs for more information. [ 563.316398] env[61906]: ERROR nova.compute.manager [ 563.316398] env[61906]: Traceback (most recent call last): [ 563.316398] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 563.316398] env[61906]: listener.cb(fileno) [ 563.316398] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 563.316398] env[61906]: result = function(*args, **kwargs) [ 563.316398] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 563.316398] env[61906]: return func(*args, **kwargs) [ 563.316398] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 563.316398] env[61906]: raise e [ 563.316398] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.316398] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 563.316398] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.316398] env[61906]: created_port_ids = self._update_ports_for_instance( [ 563.316398] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.316398] env[61906]: with excutils.save_and_reraise_exception(): [ 563.316398] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.316398] env[61906]: self.force_reraise() [ 563.316398] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.316398] env[61906]: raise self.value [ 563.316398] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.316398] env[61906]: updated_port = self._update_port( [ 563.316398] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.316398] env[61906]: _ensure_no_port_binding_failure(port) [ 563.316398] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.316398] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 563.317262] env[61906]: nova.exception.PortBindingFailed: Binding failed for port d49dae9f-3dfd-40e8-8835-ec36ed215ed8, please check neutron logs for more information. [ 563.317262] env[61906]: Removing descriptor: 17 [ 563.318040] env[61906]: ERROR nova.compute.manager [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d49dae9f-3dfd-40e8-8835-ec36ed215ed8, please check neutron logs for more information. [ 563.318040] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Traceback (most recent call last): [ 563.318040] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 563.318040] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] yield resources [ 563.318040] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 563.318040] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] self.driver.spawn(context, instance, image_meta, [ 563.318040] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 563.318040] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 563.318040] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 563.318040] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] vm_ref = self.build_virtual_machine(instance, [ 563.318040] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 563.318502] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] vif_infos = vmwarevif.get_vif_info(self._session, [ 563.318502] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 563.318502] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] for vif in network_info: [ 563.318502] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 563.318502] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] return self._sync_wrapper(fn, *args, **kwargs) [ 563.318502] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 563.318502] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] self.wait() [ 563.318502] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 563.318502] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] self[:] = self._gt.wait() [ 563.318502] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 563.318502] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] return self._exit_event.wait() [ 563.318502] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 563.318502] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] result = hub.switch() [ 563.319404] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 563.319404] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] return self.greenlet.switch() [ 563.319404] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 563.319404] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] result = function(*args, **kwargs) [ 563.319404] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 563.319404] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] return func(*args, **kwargs) [ 563.319404] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 563.319404] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] raise e [ 563.319404] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.319404] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] nwinfo = self.network_api.allocate_for_instance( [ 563.319404] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.319404] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] created_port_ids = self._update_ports_for_instance( [ 563.319404] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.319775] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] with excutils.save_and_reraise_exception(): [ 563.319775] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.319775] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] self.force_reraise() [ 563.319775] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.319775] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] raise self.value [ 563.319775] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.319775] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] updated_port = self._update_port( [ 563.319775] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.319775] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] _ensure_no_port_binding_failure(port) [ 563.319775] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.319775] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] raise exception.PortBindingFailed(port_id=port['id']) [ 563.319775] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] nova.exception.PortBindingFailed: Binding failed for port d49dae9f-3dfd-40e8-8835-ec36ed215ed8, please check neutron logs for more information. [ 563.319775] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] [ 563.320138] env[61906]: INFO nova.compute.manager [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Terminating instance [ 563.322183] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "refresh_cache-7044fd5f-a202-45a4-a4fb-73d20ef87cb5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.322183] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquired lock "refresh_cache-7044fd5f-a202-45a4-a4fb-73d20ef87cb5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.322505] env[61906]: DEBUG nova.network.neutron [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 563.482878] env[61906]: DEBUG nova.compute.manager [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 563.665062] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7580ff2f-5f8f-4b79-8b86-54f27af2b5f9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.673689] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b0588f7-bb40-4818-a469-0c4846523139 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.708493] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86dce1de-6ffd-4fd8-bb79-63fc7a43d1d8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.716514] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc68b4d-2ce7-4c58-a978-fb33a15257de {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.731012] env[61906]: DEBUG nova.compute.provider_tree [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 563.767523] env[61906]: ERROR nova.compute.manager [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 23b80c78-a492-49a8-a9e9-2d6e03a611fa, please check neutron logs for more information. [ 563.767523] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 563.767523] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.767523] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 563.767523] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.767523] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 563.767523] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.767523] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 563.767523] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.767523] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 563.767523] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.767523] env[61906]: ERROR nova.compute.manager raise self.value [ 563.767523] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.767523] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 563.767523] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.767523] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 563.769383] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.769383] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 563.769383] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 23b80c78-a492-49a8-a9e9-2d6e03a611fa, please check neutron logs for more information. [ 563.769383] env[61906]: ERROR nova.compute.manager [ 563.769383] env[61906]: Traceback (most recent call last): [ 563.769383] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 563.769383] env[61906]: listener.cb(fileno) [ 563.769383] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 563.769383] env[61906]: result = function(*args, **kwargs) [ 563.769383] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 563.769383] env[61906]: return func(*args, **kwargs) [ 563.769383] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 563.769383] env[61906]: raise e [ 563.769383] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.769383] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 563.769383] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.769383] env[61906]: created_port_ids = self._update_ports_for_instance( [ 563.769383] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.769383] env[61906]: with excutils.save_and_reraise_exception(): [ 563.769383] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.769383] env[61906]: self.force_reraise() [ 563.769383] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.769383] env[61906]: raise self.value [ 563.769383] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.769383] env[61906]: updated_port = self._update_port( [ 563.769383] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.769383] env[61906]: _ensure_no_port_binding_failure(port) [ 563.769383] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.769383] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 563.770145] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 23b80c78-a492-49a8-a9e9-2d6e03a611fa, please check neutron logs for more information. [ 563.770145] env[61906]: Removing descriptor: 16 [ 563.770145] env[61906]: ERROR nova.compute.manager [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 23b80c78-a492-49a8-a9e9-2d6e03a611fa, please check neutron logs for more information. [ 563.770145] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Traceback (most recent call last): [ 563.770145] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 563.770145] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] yield resources [ 563.770145] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 563.770145] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] self.driver.spawn(context, instance, image_meta, [ 563.770145] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 563.770145] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 563.770145] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 563.770145] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] vm_ref = self.build_virtual_machine(instance, [ 563.770445] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 563.770445] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] vif_infos = vmwarevif.get_vif_info(self._session, [ 563.770445] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 563.770445] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] for vif in network_info: [ 563.770445] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 563.770445] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] return self._sync_wrapper(fn, *args, **kwargs) [ 563.770445] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 563.770445] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] self.wait() [ 563.770445] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 563.770445] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] self[:] = self._gt.wait() [ 563.770445] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 563.770445] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] return self._exit_event.wait() [ 563.770445] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 563.770765] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] result = hub.switch() [ 563.770765] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 563.770765] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] return self.greenlet.switch() [ 563.770765] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 563.770765] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] result = function(*args, **kwargs) [ 563.770765] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 563.770765] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] return func(*args, **kwargs) [ 563.770765] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 563.770765] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] raise e [ 563.770765] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.770765] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] nwinfo = self.network_api.allocate_for_instance( [ 563.770765] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.770765] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] created_port_ids = self._update_ports_for_instance( [ 563.771077] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.771077] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] with excutils.save_and_reraise_exception(): [ 563.771077] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.771077] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] self.force_reraise() [ 563.771077] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.771077] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] raise self.value [ 563.771077] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.771077] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] updated_port = self._update_port( [ 563.771077] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.771077] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] _ensure_no_port_binding_failure(port) [ 563.771077] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.771077] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] raise exception.PortBindingFailed(port_id=port['id']) [ 563.771397] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] nova.exception.PortBindingFailed: Binding failed for port 23b80c78-a492-49a8-a9e9-2d6e03a611fa, please check neutron logs for more information. [ 563.771397] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] [ 563.771397] env[61906]: INFO nova.compute.manager [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Terminating instance [ 563.777353] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Acquiring lock "refresh_cache-1ac58495-75e5-4bce-8cb9-ecf18e0e436b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.777353] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Acquired lock "refresh_cache-1ac58495-75e5-4bce-8cb9-ecf18e0e436b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.777353] env[61906]: DEBUG nova.network.neutron [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 563.874074] env[61906]: DEBUG nova.network.neutron [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Successfully created port: 3707cc06-d460-4cf9-b352-3c7dd250ea05 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 563.877878] env[61906]: DEBUG nova.network.neutron [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.031433] env[61906]: ERROR nova.compute.manager [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9738c91a-842b-4cad-9068-a96785f62fd1, please check neutron logs for more information. [ 564.031433] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 564.031433] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 564.031433] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 564.031433] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 564.031433] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 564.031433] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 564.031433] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 564.031433] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 564.031433] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 564.031433] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 564.031433] env[61906]: ERROR nova.compute.manager raise self.value [ 564.031433] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 564.031433] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 564.031433] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 564.031433] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 564.032166] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 564.032166] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 564.032166] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9738c91a-842b-4cad-9068-a96785f62fd1, please check neutron logs for more information. [ 564.032166] env[61906]: ERROR nova.compute.manager [ 564.032166] env[61906]: Traceback (most recent call last): [ 564.032166] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 564.032166] env[61906]: listener.cb(fileno) [ 564.032166] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 564.032166] env[61906]: result = function(*args, **kwargs) [ 564.032166] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 564.032166] env[61906]: return func(*args, **kwargs) [ 564.032166] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 564.032166] env[61906]: raise e [ 564.032166] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 564.032166] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 564.032166] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 564.032166] env[61906]: created_port_ids = self._update_ports_for_instance( [ 564.032166] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 564.032166] env[61906]: with excutils.save_and_reraise_exception(): [ 564.032166] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 564.032166] env[61906]: self.force_reraise() [ 564.032166] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 564.032166] env[61906]: raise self.value [ 564.032166] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 564.032166] env[61906]: updated_port = self._update_port( [ 564.032166] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 564.032166] env[61906]: _ensure_no_port_binding_failure(port) [ 564.032166] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 564.032166] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 564.032921] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 9738c91a-842b-4cad-9068-a96785f62fd1, please check neutron logs for more information. [ 564.032921] env[61906]: Removing descriptor: 15 [ 564.032921] env[61906]: ERROR nova.compute.manager [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9738c91a-842b-4cad-9068-a96785f62fd1, please check neutron logs for more information. [ 564.032921] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] Traceback (most recent call last): [ 564.032921] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 564.032921] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] yield resources [ 564.032921] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 564.032921] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] self.driver.spawn(context, instance, image_meta, [ 564.032921] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 564.032921] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] self._vmops.spawn(context, instance, image_meta, injected_files, [ 564.032921] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 564.032921] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] vm_ref = self.build_virtual_machine(instance, [ 564.035021] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 564.035021] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] vif_infos = vmwarevif.get_vif_info(self._session, [ 564.035021] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 564.035021] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] for vif in network_info: [ 564.035021] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 564.035021] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] return self._sync_wrapper(fn, *args, **kwargs) [ 564.035021] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 564.035021] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] self.wait() [ 564.035021] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 564.035021] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] self[:] = self._gt.wait() [ 564.035021] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 564.035021] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] return self._exit_event.wait() [ 564.035021] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 564.035573] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] result = hub.switch() [ 564.035573] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 564.035573] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] return self.greenlet.switch() [ 564.035573] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 564.035573] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] result = function(*args, **kwargs) [ 564.035573] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 564.035573] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] return func(*args, **kwargs) [ 564.035573] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 564.035573] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] raise e [ 564.035573] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 564.035573] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] nwinfo = self.network_api.allocate_for_instance( [ 564.035573] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 564.035573] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] created_port_ids = self._update_ports_for_instance( [ 564.036918] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 564.036918] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] with excutils.save_and_reraise_exception(): [ 564.036918] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 564.036918] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] self.force_reraise() [ 564.036918] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 564.036918] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] raise self.value [ 564.036918] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 564.036918] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] updated_port = self._update_port( [ 564.036918] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 564.036918] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] _ensure_no_port_binding_failure(port) [ 564.036918] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 564.036918] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] raise exception.PortBindingFailed(port_id=port['id']) [ 564.037365] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] nova.exception.PortBindingFailed: Binding failed for port 9738c91a-842b-4cad-9068-a96785f62fd1, please check neutron logs for more information. [ 564.037365] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] [ 564.037365] env[61906]: INFO nova.compute.manager [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Terminating instance [ 564.037365] env[61906]: DEBUG oslo_concurrency.lockutils [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Acquiring lock "refresh_cache-739af291-794c-491d-b579-2c50c5a14785" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 564.037365] env[61906]: DEBUG oslo_concurrency.lockutils [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Acquired lock "refresh_cache-739af291-794c-491d-b579-2c50c5a14785" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 564.037365] env[61906]: DEBUG nova.network.neutron [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 564.068995] env[61906]: DEBUG nova.network.neutron [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.234174] env[61906]: DEBUG nova.scheduler.client.report [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 564.300614] env[61906]: DEBUG nova.network.neutron [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.434434] env[61906]: DEBUG nova.network.neutron [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.501504] env[61906]: DEBUG nova.compute.manager [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 564.560863] env[61906]: DEBUG nova.virt.hardware [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 564.560863] env[61906]: DEBUG nova.virt.hardware [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 564.560863] env[61906]: DEBUG nova.virt.hardware [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 564.560863] env[61906]: DEBUG nova.virt.hardware [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 564.561148] env[61906]: DEBUG nova.virt.hardware [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 564.562432] env[61906]: DEBUG nova.virt.hardware [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 564.562709] env[61906]: DEBUG nova.virt.hardware [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 564.562873] env[61906]: DEBUG nova.virt.hardware [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 564.563046] env[61906]: DEBUG nova.virt.hardware [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 564.564218] env[61906]: DEBUG nova.virt.hardware [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 564.564218] env[61906]: DEBUG nova.virt.hardware [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 564.564398] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e004fcc3-b251-480e-9b88-73a611c24661 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.573362] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Releasing lock "refresh_cache-7044fd5f-a202-45a4-a4fb-73d20ef87cb5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 564.573362] env[61906]: DEBUG nova.compute.manager [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 564.575611] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 564.575940] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-39d05ddb-a9c4-4676-9d51-b15c793a167e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.582663] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa003524-0b20-439d-b2d9-8851ffe3cbc8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.589066] env[61906]: DEBUG nova.network.neutron [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.603446] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5006132-27dc-40b2-83f3-980de91459f6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.630607] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Acquiring lock "da9ac1bd-97c8-4c05-9364-8bd60be50ff2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.630607] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Lock "da9ac1bd-97c8-4c05-9364-8bd60be50ff2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.639125] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7044fd5f-a202-45a4-a4fb-73d20ef87cb5 could not be found. [ 564.639988] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 564.639988] env[61906]: INFO nova.compute.manager [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Took 0.07 seconds to destroy the instance on the hypervisor. [ 564.639988] env[61906]: DEBUG oslo.service.loopingcall [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 564.640238] env[61906]: DEBUG nova.compute.manager [-] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 564.640922] env[61906]: DEBUG nova.network.neutron [-] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 564.678243] env[61906]: DEBUG nova.network.neutron [-] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.742912] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.280s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.743485] env[61906]: DEBUG nova.compute.manager [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 564.747695] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 4.113s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.896135] env[61906]: DEBUG nova.network.neutron [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.938670] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Releasing lock "refresh_cache-1ac58495-75e5-4bce-8cb9-ecf18e0e436b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 564.939171] env[61906]: DEBUG nova.compute.manager [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 564.939322] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 564.939693] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-84b185b2-7c4b-41b9-9755-2e3602cb8e70 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.954957] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8495b628-50ed-4d94-8a2a-930bdaebc3d8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.983909] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1ac58495-75e5-4bce-8cb9-ecf18e0e436b could not be found. [ 564.987045] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 564.987045] env[61906]: INFO nova.compute.manager [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Took 0.05 seconds to destroy the instance on the hypervisor. [ 564.987045] env[61906]: DEBUG oslo.service.loopingcall [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 564.987045] env[61906]: DEBUG nova.compute.manager [-] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 564.987045] env[61906]: DEBUG nova.network.neutron [-] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 564.992362] env[61906]: DEBUG oslo_concurrency.lockutils [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Acquiring lock "eb4cac2e-85d8-46b6-a4d4-ff12ee70e435" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.992582] env[61906]: DEBUG oslo_concurrency.lockutils [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Lock "eb4cac2e-85d8-46b6-a4d4-ff12ee70e435" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.008830] env[61906]: DEBUG nova.network.neutron [-] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 565.135070] env[61906]: DEBUG nova.compute.manager [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 565.182946] env[61906]: DEBUG nova.network.neutron [-] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.259143] env[61906]: DEBUG nova.compute.utils [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 565.261081] env[61906]: DEBUG nova.compute.manager [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 565.261081] env[61906]: DEBUG nova.network.neutron [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 565.322529] env[61906]: DEBUG nova.compute.manager [req-7670b99b-68df-4888-9bc3-8c473cd8fbce req-1454e508-01da-48a7-9bbb-22dca21711a9 service nova] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Received event network-vif-deleted-fee6c75a-e766-4956-a3e2-1149ebf95e92 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 565.322529] env[61906]: DEBUG nova.compute.manager [req-7670b99b-68df-4888-9bc3-8c473cd8fbce req-1454e508-01da-48a7-9bbb-22dca21711a9 service nova] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Received event network-changed-23b80c78-a492-49a8-a9e9-2d6e03a611fa {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 565.322529] env[61906]: DEBUG nova.compute.manager [req-7670b99b-68df-4888-9bc3-8c473cd8fbce req-1454e508-01da-48a7-9bbb-22dca21711a9 service nova] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Refreshing instance network info cache due to event network-changed-23b80c78-a492-49a8-a9e9-2d6e03a611fa. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 565.322529] env[61906]: DEBUG oslo_concurrency.lockutils [req-7670b99b-68df-4888-9bc3-8c473cd8fbce req-1454e508-01da-48a7-9bbb-22dca21711a9 service nova] Acquiring lock "refresh_cache-1ac58495-75e5-4bce-8cb9-ecf18e0e436b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.322529] env[61906]: DEBUG oslo_concurrency.lockutils [req-7670b99b-68df-4888-9bc3-8c473cd8fbce req-1454e508-01da-48a7-9bbb-22dca21711a9 service nova] Acquired lock "refresh_cache-1ac58495-75e5-4bce-8cb9-ecf18e0e436b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.322706] env[61906]: DEBUG nova.network.neutron [req-7670b99b-68df-4888-9bc3-8c473cd8fbce req-1454e508-01da-48a7-9bbb-22dca21711a9 service nova] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Refreshing network info cache for port 23b80c78-a492-49a8-a9e9-2d6e03a611fa {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 565.399872] env[61906]: DEBUG oslo_concurrency.lockutils [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Releasing lock "refresh_cache-739af291-794c-491d-b579-2c50c5a14785" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 565.401868] env[61906]: DEBUG nova.compute.manager [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 565.401868] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 565.401868] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8b6934a4-3595-4263-9bf2-4c6fe5b76983 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.411920] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce6bec90-3c18-4475-9d41-9da438ed2905 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.438411] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 739af291-794c-491d-b579-2c50c5a14785 could not be found. [ 565.438678] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 565.438865] env[61906]: INFO nova.compute.manager [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Took 0.04 seconds to destroy the instance on the hypervisor. [ 565.439379] env[61906]: DEBUG oslo.service.loopingcall [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 565.442474] env[61906]: DEBUG nova.compute.manager [-] [instance: 739af291-794c-491d-b579-2c50c5a14785] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 565.442474] env[61906]: DEBUG nova.network.neutron [-] [instance: 739af291-794c-491d-b579-2c50c5a14785] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 565.445099] env[61906]: DEBUG nova.policy [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '463262c2e40e4db4a4469aaa5e6ac708', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '03f866963a584575a4444b1f0493cd71', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 565.482243] env[61906]: DEBUG nova.network.neutron [-] [instance: 739af291-794c-491d-b579-2c50c5a14785] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 565.492632] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f154de7-4798-42b5-9831-bef872c5132e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.496390] env[61906]: DEBUG nova.compute.manager [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 565.505312] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f2ba460-b231-439c-84c1-a3a9b60c6249 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.511460] env[61906]: DEBUG nova.network.neutron [-] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.543646] env[61906]: INFO nova.compute.manager [-] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Took 0.56 seconds to deallocate network for instance. [ 565.548226] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ab11e94-f714-420f-88a8-1be19bfb4608 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.553232] env[61906]: DEBUG nova.compute.claims [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 565.553232] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.558864] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f2797c6-efd3-4f6b-8d04-34df56520d5c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.578945] env[61906]: DEBUG nova.compute.provider_tree [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 565.666410] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.687708] env[61906]: INFO nova.compute.manager [-] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Took 1.05 seconds to deallocate network for instance. [ 565.690274] env[61906]: DEBUG nova.compute.claims [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 565.690477] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.767555] env[61906]: DEBUG nova.compute.manager [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 565.891760] env[61906]: DEBUG nova.network.neutron [req-7670b99b-68df-4888-9bc3-8c473cd8fbce req-1454e508-01da-48a7-9bbb-22dca21711a9 service nova] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 565.954881] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Acquiring lock "263da5ea-892b-4142-b487-96444abc97be" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.955312] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Lock "263da5ea-892b-4142-b487-96444abc97be" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.984869] env[61906]: DEBUG nova.network.neutron [-] [instance: 739af291-794c-491d-b579-2c50c5a14785] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.029439] env[61906]: DEBUG nova.network.neutron [req-7670b99b-68df-4888-9bc3-8c473cd8fbce req-1454e508-01da-48a7-9bbb-22dca21711a9 service nova] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.033804] env[61906]: DEBUG oslo_concurrency.lockutils [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.059034] env[61906]: DEBUG nova.compute.manager [req-88a57f37-7941-4dbc-8b11-a85b5a49390c req-a504a8a0-703b-4b04-b82d-134f6c776b3c service nova] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Received event network-changed-d49dae9f-3dfd-40e8-8835-ec36ed215ed8 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 566.060028] env[61906]: DEBUG nova.compute.manager [req-88a57f37-7941-4dbc-8b11-a85b5a49390c req-a504a8a0-703b-4b04-b82d-134f6c776b3c service nova] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Refreshing instance network info cache due to event network-changed-d49dae9f-3dfd-40e8-8835-ec36ed215ed8. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 566.060028] env[61906]: DEBUG oslo_concurrency.lockutils [req-88a57f37-7941-4dbc-8b11-a85b5a49390c req-a504a8a0-703b-4b04-b82d-134f6c776b3c service nova] Acquiring lock "refresh_cache-7044fd5f-a202-45a4-a4fb-73d20ef87cb5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.060028] env[61906]: DEBUG oslo_concurrency.lockutils [req-88a57f37-7941-4dbc-8b11-a85b5a49390c req-a504a8a0-703b-4b04-b82d-134f6c776b3c service nova] Acquired lock "refresh_cache-7044fd5f-a202-45a4-a4fb-73d20ef87cb5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.060028] env[61906]: DEBUG nova.network.neutron [req-88a57f37-7941-4dbc-8b11-a85b5a49390c req-a504a8a0-703b-4b04-b82d-134f6c776b3c service nova] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Refreshing network info cache for port d49dae9f-3dfd-40e8-8835-ec36ed215ed8 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 566.082637] env[61906]: DEBUG nova.scheduler.client.report [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 566.415899] env[61906]: DEBUG nova.network.neutron [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Successfully created port: 0fb5cdf6-dbaf-4b77-aaf7-19f1195c7614 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 566.491617] env[61906]: INFO nova.compute.manager [-] [instance: 739af291-794c-491d-b579-2c50c5a14785] Took 1.05 seconds to deallocate network for instance. [ 566.494924] env[61906]: DEBUG nova.compute.claims [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 566.495221] env[61906]: DEBUG oslo_concurrency.lockutils [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.533130] env[61906]: DEBUG oslo_concurrency.lockutils [req-7670b99b-68df-4888-9bc3-8c473cd8fbce req-1454e508-01da-48a7-9bbb-22dca21711a9 service nova] Releasing lock "refresh_cache-1ac58495-75e5-4bce-8cb9-ecf18e0e436b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.533406] env[61906]: DEBUG nova.compute.manager [req-7670b99b-68df-4888-9bc3-8c473cd8fbce req-1454e508-01da-48a7-9bbb-22dca21711a9 service nova] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Received event network-vif-deleted-23b80c78-a492-49a8-a9e9-2d6e03a611fa {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 566.591182] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.841s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.591182] env[61906]: ERROR nova.compute.manager [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8a65f04d-2ad4-4d14-a8f2-0f2200525b06, please check neutron logs for more information. [ 566.591182] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] Traceback (most recent call last): [ 566.591182] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 566.591182] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] self.driver.spawn(context, instance, image_meta, [ 566.591182] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 566.591182] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 566.591182] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 566.591182] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] vm_ref = self.build_virtual_machine(instance, [ 566.591493] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 566.591493] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] vif_infos = vmwarevif.get_vif_info(self._session, [ 566.591493] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 566.591493] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] for vif in network_info: [ 566.591493] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 566.591493] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] return self._sync_wrapper(fn, *args, **kwargs) [ 566.591493] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 566.591493] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] self.wait() [ 566.591493] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 566.591493] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] self[:] = self._gt.wait() [ 566.591493] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 566.591493] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] return self._exit_event.wait() [ 566.591493] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 566.591837] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] result = hub.switch() [ 566.591837] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 566.591837] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] return self.greenlet.switch() [ 566.591837] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 566.591837] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] result = function(*args, **kwargs) [ 566.591837] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 566.591837] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] return func(*args, **kwargs) [ 566.591837] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 566.591837] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] raise e [ 566.591837] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 566.591837] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] nwinfo = self.network_api.allocate_for_instance( [ 566.591837] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 566.591837] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] created_port_ids = self._update_ports_for_instance( [ 566.592183] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 566.592183] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] with excutils.save_and_reraise_exception(): [ 566.592183] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 566.592183] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] self.force_reraise() [ 566.592183] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 566.592183] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] raise self.value [ 566.592183] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 566.592183] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] updated_port = self._update_port( [ 566.592183] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 566.592183] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] _ensure_no_port_binding_failure(port) [ 566.592183] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 566.592183] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] raise exception.PortBindingFailed(port_id=port['id']) [ 566.592539] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] nova.exception.PortBindingFailed: Binding failed for port 8a65f04d-2ad4-4d14-a8f2-0f2200525b06, please check neutron logs for more information. [ 566.592539] env[61906]: ERROR nova.compute.manager [instance: 47f866b0-5752-4439-9dca-de44934955a6] [ 566.592539] env[61906]: DEBUG nova.compute.utils [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Binding failed for port 8a65f04d-2ad4-4d14-a8f2-0f2200525b06, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 566.595803] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 4.010s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.600264] env[61906]: DEBUG nova.compute.manager [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Build of instance 47f866b0-5752-4439-9dca-de44934955a6 was re-scheduled: Binding failed for port 8a65f04d-2ad4-4d14-a8f2-0f2200525b06, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 566.600788] env[61906]: DEBUG nova.compute.manager [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 566.601066] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Acquiring lock "refresh_cache-47f866b0-5752-4439-9dca-de44934955a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.605042] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Acquired lock "refresh_cache-47f866b0-5752-4439-9dca-de44934955a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.605042] env[61906]: DEBUG nova.network.neutron [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 566.606566] env[61906]: DEBUG nova.network.neutron [req-88a57f37-7941-4dbc-8b11-a85b5a49390c req-a504a8a0-703b-4b04-b82d-134f6c776b3c service nova] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 566.725102] env[61906]: ERROR nova.compute.manager [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3707cc06-d460-4cf9-b352-3c7dd250ea05, please check neutron logs for more information. [ 566.725102] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 566.725102] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 566.725102] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 566.725102] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 566.725102] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 566.725102] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 566.725102] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 566.725102] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 566.725102] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 566.725102] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 566.725102] env[61906]: ERROR nova.compute.manager raise self.value [ 566.725102] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 566.725102] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 566.725102] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 566.725102] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 566.725603] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 566.725603] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 566.725603] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3707cc06-d460-4cf9-b352-3c7dd250ea05, please check neutron logs for more information. [ 566.725603] env[61906]: ERROR nova.compute.manager [ 566.725603] env[61906]: Traceback (most recent call last): [ 566.725603] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 566.725603] env[61906]: listener.cb(fileno) [ 566.725603] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 566.725603] env[61906]: result = function(*args, **kwargs) [ 566.725603] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 566.725603] env[61906]: return func(*args, **kwargs) [ 566.725603] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 566.725603] env[61906]: raise e [ 566.725603] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 566.725603] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 566.725603] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 566.725603] env[61906]: created_port_ids = self._update_ports_for_instance( [ 566.725603] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 566.725603] env[61906]: with excutils.save_and_reraise_exception(): [ 566.725603] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 566.725603] env[61906]: self.force_reraise() [ 566.725603] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 566.725603] env[61906]: raise self.value [ 566.725603] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 566.725603] env[61906]: updated_port = self._update_port( [ 566.725603] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 566.725603] env[61906]: _ensure_no_port_binding_failure(port) [ 566.725603] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 566.725603] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 566.726398] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 3707cc06-d460-4cf9-b352-3c7dd250ea05, please check neutron logs for more information. [ 566.726398] env[61906]: Removing descriptor: 18 [ 566.726398] env[61906]: ERROR nova.compute.manager [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3707cc06-d460-4cf9-b352-3c7dd250ea05, please check neutron logs for more information. [ 566.726398] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] Traceback (most recent call last): [ 566.726398] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 566.726398] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] yield resources [ 566.726398] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 566.726398] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] self.driver.spawn(context, instance, image_meta, [ 566.726398] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 566.726398] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 566.726398] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 566.726398] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] vm_ref = self.build_virtual_machine(instance, [ 566.726776] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 566.726776] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] vif_infos = vmwarevif.get_vif_info(self._session, [ 566.726776] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 566.726776] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] for vif in network_info: [ 566.726776] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 566.726776] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] return self._sync_wrapper(fn, *args, **kwargs) [ 566.726776] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 566.726776] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] self.wait() [ 566.726776] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 566.726776] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] self[:] = self._gt.wait() [ 566.726776] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 566.726776] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] return self._exit_event.wait() [ 566.726776] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 566.727199] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] result = hub.switch() [ 566.727199] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 566.727199] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] return self.greenlet.switch() [ 566.727199] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 566.727199] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] result = function(*args, **kwargs) [ 566.727199] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 566.727199] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] return func(*args, **kwargs) [ 566.727199] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 566.727199] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] raise e [ 566.727199] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 566.727199] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] nwinfo = self.network_api.allocate_for_instance( [ 566.727199] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 566.727199] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] created_port_ids = self._update_ports_for_instance( [ 566.727584] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 566.727584] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] with excutils.save_and_reraise_exception(): [ 566.727584] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 566.727584] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] self.force_reraise() [ 566.727584] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 566.727584] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] raise self.value [ 566.727584] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 566.727584] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] updated_port = self._update_port( [ 566.727584] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 566.727584] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] _ensure_no_port_binding_failure(port) [ 566.727584] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 566.727584] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] raise exception.PortBindingFailed(port_id=port['id']) [ 566.727929] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] nova.exception.PortBindingFailed: Binding failed for port 3707cc06-d460-4cf9-b352-3c7dd250ea05, please check neutron logs for more information. [ 566.727929] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] [ 566.727929] env[61906]: INFO nova.compute.manager [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Terminating instance [ 566.731588] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Acquiring lock "refresh_cache-bf480958-27a3-4450-b09c-cba96375712b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.731743] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Acquired lock "refresh_cache-bf480958-27a3-4450-b09c-cba96375712b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.732433] env[61906]: DEBUG nova.network.neutron [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 566.780728] env[61906]: DEBUG nova.compute.manager [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 566.819263] env[61906]: DEBUG nova.virt.hardware [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 566.819722] env[61906]: DEBUG nova.virt.hardware [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 566.820183] env[61906]: DEBUG nova.virt.hardware [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 566.820923] env[61906]: DEBUG nova.virt.hardware [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 566.820923] env[61906]: DEBUG nova.virt.hardware [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 566.820923] env[61906]: DEBUG nova.virt.hardware [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 566.821057] env[61906]: DEBUG nova.virt.hardware [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 566.821163] env[61906]: DEBUG nova.virt.hardware [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 566.821318] env[61906]: DEBUG nova.virt.hardware [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 566.821756] env[61906]: DEBUG nova.virt.hardware [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 566.822095] env[61906]: DEBUG nova.virt.hardware [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 566.822937] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1450088f-b888-478c-9139-abc70f9d19c2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.835265] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e98d9f71-26d8-4c1d-89be-d25826ac73da {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.947912] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Acquiring lock "1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.948158] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Lock "1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.007809] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Acquiring lock "318a3224-a156-4354-a73b-460168c0eb9c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.009355] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Lock "318a3224-a156-4354-a73b-460168c0eb9c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.080458] env[61906]: DEBUG nova.network.neutron [req-88a57f37-7941-4dbc-8b11-a85b5a49390c req-a504a8a0-703b-4b04-b82d-134f6c776b3c service nova] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.148438] env[61906]: DEBUG nova.network.neutron [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 567.269174] env[61906]: DEBUG nova.network.neutron [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 567.275577] env[61906]: DEBUG nova.network.neutron [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.459526] env[61906]: DEBUG nova.network.neutron [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.572928] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Acquiring lock "7ed5b09c-5ba1-479c-96b4-75924df84e75" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.573223] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Lock "7ed5b09c-5ba1-479c-96b4-75924df84e75" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.584183] env[61906]: DEBUG oslo_concurrency.lockutils [req-88a57f37-7941-4dbc-8b11-a85b5a49390c req-a504a8a0-703b-4b04-b82d-134f6c776b3c service nova] Releasing lock "refresh_cache-7044fd5f-a202-45a4-a4fb-73d20ef87cb5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.584914] env[61906]: DEBUG nova.compute.manager [req-88a57f37-7941-4dbc-8b11-a85b5a49390c req-a504a8a0-703b-4b04-b82d-134f6c776b3c service nova] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Received event network-vif-deleted-d49dae9f-3dfd-40e8-8835-ec36ed215ed8 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 567.584914] env[61906]: DEBUG nova.compute.manager [req-88a57f37-7941-4dbc-8b11-a85b5a49390c req-a504a8a0-703b-4b04-b82d-134f6c776b3c service nova] [instance: 739af291-794c-491d-b579-2c50c5a14785] Received event network-changed-9738c91a-842b-4cad-9068-a96785f62fd1 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 567.584914] env[61906]: DEBUG nova.compute.manager [req-88a57f37-7941-4dbc-8b11-a85b5a49390c req-a504a8a0-703b-4b04-b82d-134f6c776b3c service nova] [instance: 739af291-794c-491d-b579-2c50c5a14785] Refreshing instance network info cache due to event network-changed-9738c91a-842b-4cad-9068-a96785f62fd1. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 567.584914] env[61906]: DEBUG oslo_concurrency.lockutils [req-88a57f37-7941-4dbc-8b11-a85b5a49390c req-a504a8a0-703b-4b04-b82d-134f6c776b3c service nova] Acquiring lock "refresh_cache-739af291-794c-491d-b579-2c50c5a14785" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.585108] env[61906]: DEBUG oslo_concurrency.lockutils [req-88a57f37-7941-4dbc-8b11-a85b5a49390c req-a504a8a0-703b-4b04-b82d-134f6c776b3c service nova] Acquired lock "refresh_cache-739af291-794c-491d-b579-2c50c5a14785" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.585157] env[61906]: DEBUG nova.network.neutron [req-88a57f37-7941-4dbc-8b11-a85b5a49390c req-a504a8a0-703b-4b04-b82d-134f6c776b3c service nova] [instance: 739af291-794c-491d-b579-2c50c5a14785] Refreshing network info cache for port 9738c91a-842b-4cad-9068-a96785f62fd1 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 567.778642] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Releasing lock "refresh_cache-47f866b0-5752-4439-9dca-de44934955a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.778893] env[61906]: DEBUG nova.compute.manager [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 567.779410] env[61906]: DEBUG nova.compute.manager [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 567.779410] env[61906]: DEBUG nova.network.neutron [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 567.808030] env[61906]: DEBUG nova.network.neutron [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 567.963841] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Releasing lock "refresh_cache-bf480958-27a3-4450-b09c-cba96375712b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.964349] env[61906]: DEBUG nova.compute.manager [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 567.964494] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 567.964790] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cb6aa230-7c84-4be6-9441-69216a3ff48c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.978294] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-613a0e08-dc87-4dd5-abd4-f2eeb4b6667c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.006317] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bf480958-27a3-4450-b09c-cba96375712b could not be found. [ 568.007456] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 568.007456] env[61906]: INFO nova.compute.manager [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 568.007456] env[61906]: DEBUG oslo.service.loopingcall [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 568.007456] env[61906]: DEBUG nova.compute.manager [-] [instance: bf480958-27a3-4450-b09c-cba96375712b] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 568.007456] env[61906]: DEBUG nova.network.neutron [-] [instance: bf480958-27a3-4450-b09c-cba96375712b] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 568.038104] env[61906]: DEBUG nova.network.neutron [-] [instance: bf480958-27a3-4450-b09c-cba96375712b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 568.109860] env[61906]: DEBUG nova.network.neutron [req-88a57f37-7941-4dbc-8b11-a85b5a49390c req-a504a8a0-703b-4b04-b82d-134f6c776b3c service nova] [instance: 739af291-794c-491d-b579-2c50c5a14785] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 568.139171] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 47f866b0-5752-4439-9dca-de44934955a6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 568.139171] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 7044fd5f-a202-45a4-a4fb-73d20ef87cb5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 568.139300] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 402707da-f091-40fb-b5c7-3a2b1429ecce actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 568.139562] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 739af291-794c-491d-b579-2c50c5a14785 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 568.139562] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 1ac58495-75e5-4bce-8cb9-ecf18e0e436b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 568.139663] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance bf480958-27a3-4450-b09c-cba96375712b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 568.139739] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 568.269010] env[61906]: DEBUG nova.network.neutron [req-88a57f37-7941-4dbc-8b11-a85b5a49390c req-a504a8a0-703b-4b04-b82d-134f6c776b3c service nova] [instance: 739af291-794c-491d-b579-2c50c5a14785] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.311023] env[61906]: DEBUG nova.network.neutron [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.545083] env[61906]: DEBUG nova.network.neutron [-] [instance: bf480958-27a3-4450-b09c-cba96375712b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.644052] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 75124494-73f0-4188-817b-1c7b4a85c8de has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 568.775739] env[61906]: DEBUG oslo_concurrency.lockutils [req-88a57f37-7941-4dbc-8b11-a85b5a49390c req-a504a8a0-703b-4b04-b82d-134f6c776b3c service nova] Releasing lock "refresh_cache-739af291-794c-491d-b579-2c50c5a14785" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.776025] env[61906]: DEBUG nova.compute.manager [req-88a57f37-7941-4dbc-8b11-a85b5a49390c req-a504a8a0-703b-4b04-b82d-134f6c776b3c service nova] [instance: 739af291-794c-491d-b579-2c50c5a14785] Received event network-vif-deleted-9738c91a-842b-4cad-9068-a96785f62fd1 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 568.776309] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Acquiring lock "a9249055-f6bf-4693-b0f7-5d8b63510a35" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.776511] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Lock "a9249055-f6bf-4693-b0f7-5d8b63510a35" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.816031] env[61906]: INFO nova.compute.manager [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 47f866b0-5752-4439-9dca-de44934955a6] Took 1.03 seconds to deallocate network for instance. [ 568.972900] env[61906]: ERROR nova.compute.manager [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0fb5cdf6-dbaf-4b77-aaf7-19f1195c7614, please check neutron logs for more information. [ 568.972900] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 568.972900] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.972900] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 568.972900] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 568.972900] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 568.972900] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 568.972900] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 568.972900] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.972900] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 568.972900] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.972900] env[61906]: ERROR nova.compute.manager raise self.value [ 568.972900] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 568.972900] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 568.972900] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.972900] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 568.973389] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.973389] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 568.973389] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0fb5cdf6-dbaf-4b77-aaf7-19f1195c7614, please check neutron logs for more information. [ 568.973389] env[61906]: ERROR nova.compute.manager [ 568.973389] env[61906]: Traceback (most recent call last): [ 568.973389] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 568.973389] env[61906]: listener.cb(fileno) [ 568.973389] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.973389] env[61906]: result = function(*args, **kwargs) [ 568.973389] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 568.973389] env[61906]: return func(*args, **kwargs) [ 568.973389] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 568.973389] env[61906]: raise e [ 568.973389] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.973389] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 568.973389] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 568.973389] env[61906]: created_port_ids = self._update_ports_for_instance( [ 568.973389] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 568.973389] env[61906]: with excutils.save_and_reraise_exception(): [ 568.973389] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.973389] env[61906]: self.force_reraise() [ 568.973389] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.973389] env[61906]: raise self.value [ 568.973389] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 568.973389] env[61906]: updated_port = self._update_port( [ 568.973389] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.973389] env[61906]: _ensure_no_port_binding_failure(port) [ 568.973389] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.973389] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 568.974181] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 0fb5cdf6-dbaf-4b77-aaf7-19f1195c7614, please check neutron logs for more information. [ 568.974181] env[61906]: Removing descriptor: 16 [ 568.974181] env[61906]: ERROR nova.compute.manager [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0fb5cdf6-dbaf-4b77-aaf7-19f1195c7614, please check neutron logs for more information. [ 568.974181] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Traceback (most recent call last): [ 568.974181] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 568.974181] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] yield resources [ 568.974181] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 568.974181] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] self.driver.spawn(context, instance, image_meta, [ 568.974181] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 568.974181] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] self._vmops.spawn(context, instance, image_meta, injected_files, [ 568.974181] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 568.974181] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] vm_ref = self.build_virtual_machine(instance, [ 568.974529] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 568.974529] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] vif_infos = vmwarevif.get_vif_info(self._session, [ 568.974529] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 568.974529] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] for vif in network_info: [ 568.974529] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 568.974529] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] return self._sync_wrapper(fn, *args, **kwargs) [ 568.974529] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 568.974529] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] self.wait() [ 568.974529] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 568.974529] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] self[:] = self._gt.wait() [ 568.974529] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 568.974529] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] return self._exit_event.wait() [ 568.974529] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 568.974905] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] result = hub.switch() [ 568.974905] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 568.974905] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] return self.greenlet.switch() [ 568.974905] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.974905] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] result = function(*args, **kwargs) [ 568.974905] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 568.974905] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] return func(*args, **kwargs) [ 568.974905] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 568.974905] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] raise e [ 568.974905] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.974905] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] nwinfo = self.network_api.allocate_for_instance( [ 568.974905] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 568.974905] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] created_port_ids = self._update_ports_for_instance( [ 568.975303] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 568.975303] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] with excutils.save_and_reraise_exception(): [ 568.975303] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.975303] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] self.force_reraise() [ 568.975303] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.975303] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] raise self.value [ 568.975303] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 568.975303] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] updated_port = self._update_port( [ 568.975303] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.975303] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] _ensure_no_port_binding_failure(port) [ 568.975303] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.975303] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] raise exception.PortBindingFailed(port_id=port['id']) [ 568.975668] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] nova.exception.PortBindingFailed: Binding failed for port 0fb5cdf6-dbaf-4b77-aaf7-19f1195c7614, please check neutron logs for more information. [ 568.975668] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] [ 568.975668] env[61906]: INFO nova.compute.manager [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Terminating instance [ 568.978620] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Acquiring lock "refresh_cache-75ea63d7-6fc7-4a9f-b3df-05c6d12fe913" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.978620] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Acquired lock "refresh_cache-75ea63d7-6fc7-4a9f-b3df-05c6d12fe913" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.978620] env[61906]: DEBUG nova.network.neutron [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 569.052631] env[61906]: INFO nova.compute.manager [-] [instance: bf480958-27a3-4450-b09c-cba96375712b] Took 1.05 seconds to deallocate network for instance. [ 569.055683] env[61906]: DEBUG nova.compute.claims [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 569.055894] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.153998] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance da9ac1bd-97c8-4c05-9364-8bd60be50ff2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 569.506306] env[61906]: DEBUG nova.network.neutron [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 569.583449] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Acquiring lock "bff5e099-f6ad-41bd-a55a-6987310c37cf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.583783] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Lock "bff5e099-f6ad-41bd-a55a-6987310c37cf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.619205] env[61906]: DEBUG nova.network.neutron [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.656134] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance eb4cac2e-85d8-46b6-a4d4-ff12ee70e435 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 569.738860] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Acquiring lock "e64791e3-7269-4953-bfe1-fe776744f493" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.739140] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Lock "e64791e3-7269-4953-bfe1-fe776744f493" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.849500] env[61906]: DEBUG nova.compute.manager [req-74dc718f-9d8e-46a8-ac32-d7cc51c770c7 req-34e6a92a-6136-40d7-9d13-74f1b690b2de service nova] [instance: bf480958-27a3-4450-b09c-cba96375712b] Received event network-changed-3707cc06-d460-4cf9-b352-3c7dd250ea05 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 569.851229] env[61906]: DEBUG nova.compute.manager [req-74dc718f-9d8e-46a8-ac32-d7cc51c770c7 req-34e6a92a-6136-40d7-9d13-74f1b690b2de service nova] [instance: bf480958-27a3-4450-b09c-cba96375712b] Refreshing instance network info cache due to event network-changed-3707cc06-d460-4cf9-b352-3c7dd250ea05. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 569.851229] env[61906]: DEBUG oslo_concurrency.lockutils [req-74dc718f-9d8e-46a8-ac32-d7cc51c770c7 req-34e6a92a-6136-40d7-9d13-74f1b690b2de service nova] Acquiring lock "refresh_cache-bf480958-27a3-4450-b09c-cba96375712b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.851229] env[61906]: DEBUG oslo_concurrency.lockutils [req-74dc718f-9d8e-46a8-ac32-d7cc51c770c7 req-34e6a92a-6136-40d7-9d13-74f1b690b2de service nova] Acquired lock "refresh_cache-bf480958-27a3-4450-b09c-cba96375712b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.851229] env[61906]: DEBUG nova.network.neutron [req-74dc718f-9d8e-46a8-ac32-d7cc51c770c7 req-34e6a92a-6136-40d7-9d13-74f1b690b2de service nova] [instance: bf480958-27a3-4450-b09c-cba96375712b] Refreshing network info cache for port 3707cc06-d460-4cf9-b352-3c7dd250ea05 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 569.864674] env[61906]: INFO nova.scheduler.client.report [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Deleted allocations for instance 47f866b0-5752-4439-9dca-de44934955a6 [ 569.879710] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Acquiring lock "9fe29bcd-a682-4973-9bb8-069adc204086" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.879937] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Lock "9fe29bcd-a682-4973-9bb8-069adc204086" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.121777] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Releasing lock "refresh_cache-75ea63d7-6fc7-4a9f-b3df-05c6d12fe913" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.122330] env[61906]: DEBUG nova.compute.manager [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 570.122601] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 570.122943] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cfbfa127-0f37-42ac-ac89-568b4784ff79 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.135162] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae9e8f7a-5360-4e80-9b0e-103f31584d9f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.162509] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 263da5ea-892b-4142-b487-96444abc97be has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 570.168038] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913 could not be found. [ 570.168038] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 570.168162] env[61906]: INFO nova.compute.manager [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Took 0.05 seconds to destroy the instance on the hypervisor. [ 570.169784] env[61906]: DEBUG oslo.service.loopingcall [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 570.169784] env[61906]: DEBUG nova.compute.manager [-] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 570.169784] env[61906]: DEBUG nova.network.neutron [-] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 570.194628] env[61906]: DEBUG nova.network.neutron [-] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 570.383244] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2036ce4d-66ad-417b-b5a7-5c5acee20c07 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Lock "47f866b0-5752-4439-9dca-de44934955a6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.571s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.384641] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "47f866b0-5752-4439-9dca-de44934955a6" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 18.917s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.384886] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-97223777-fd4c-476e-a9db-c417a8142d1e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.396412] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c47e985-afe4-44de-8d1d-c34cceaaf576 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.418274] env[61906]: DEBUG nova.network.neutron [req-74dc718f-9d8e-46a8-ac32-d7cc51c770c7 req-34e6a92a-6136-40d7-9d13-74f1b690b2de service nova] [instance: bf480958-27a3-4450-b09c-cba96375712b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 570.558731] env[61906]: DEBUG nova.network.neutron [req-74dc718f-9d8e-46a8-ac32-d7cc51c770c7 req-34e6a92a-6136-40d7-9d13-74f1b690b2de service nova] [instance: bf480958-27a3-4450-b09c-cba96375712b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.630074] env[61906]: DEBUG nova.compute.manager [req-488bb97b-d30c-43d3-a604-01cbfc9a683e req-7df1402e-2d84-4d80-8427-56fff6dd3658 service nova] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Received event network-changed-0fb5cdf6-dbaf-4b77-aaf7-19f1195c7614 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 570.630320] env[61906]: DEBUG nova.compute.manager [req-488bb97b-d30c-43d3-a604-01cbfc9a683e req-7df1402e-2d84-4d80-8427-56fff6dd3658 service nova] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Refreshing instance network info cache due to event network-changed-0fb5cdf6-dbaf-4b77-aaf7-19f1195c7614. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 570.630676] env[61906]: DEBUG oslo_concurrency.lockutils [req-488bb97b-d30c-43d3-a604-01cbfc9a683e req-7df1402e-2d84-4d80-8427-56fff6dd3658 service nova] Acquiring lock "refresh_cache-75ea63d7-6fc7-4a9f-b3df-05c6d12fe913" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 570.630877] env[61906]: DEBUG oslo_concurrency.lockutils [req-488bb97b-d30c-43d3-a604-01cbfc9a683e req-7df1402e-2d84-4d80-8427-56fff6dd3658 service nova] Acquired lock "refresh_cache-75ea63d7-6fc7-4a9f-b3df-05c6d12fe913" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 570.631673] env[61906]: DEBUG nova.network.neutron [req-488bb97b-d30c-43d3-a604-01cbfc9a683e req-7df1402e-2d84-4d80-8427-56fff6dd3658 service nova] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Refreshing network info cache for port 0fb5cdf6-dbaf-4b77-aaf7-19f1195c7614 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 570.671885] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 570.696527] env[61906]: DEBUG nova.network.neutron [-] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.889103] env[61906]: DEBUG nova.compute.manager [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 570.934895] env[61906]: INFO nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 47f866b0-5752-4439-9dca-de44934955a6] During the sync_power process the instance has moved from host None to host cpu-1 [ 570.935146] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "47f866b0-5752-4439-9dca-de44934955a6" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.551s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.062438] env[61906]: DEBUG oslo_concurrency.lockutils [req-74dc718f-9d8e-46a8-ac32-d7cc51c770c7 req-34e6a92a-6136-40d7-9d13-74f1b690b2de service nova] Releasing lock "refresh_cache-bf480958-27a3-4450-b09c-cba96375712b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.062438] env[61906]: DEBUG nova.compute.manager [req-74dc718f-9d8e-46a8-ac32-d7cc51c770c7 req-34e6a92a-6136-40d7-9d13-74f1b690b2de service nova] [instance: bf480958-27a3-4450-b09c-cba96375712b] Received event network-vif-deleted-3707cc06-d460-4cf9-b352-3c7dd250ea05 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 571.158463] env[61906]: DEBUG nova.network.neutron [req-488bb97b-d30c-43d3-a604-01cbfc9a683e req-7df1402e-2d84-4d80-8427-56fff6dd3658 service nova] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 571.177664] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 318a3224-a156-4354-a73b-460168c0eb9c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 571.200598] env[61906]: INFO nova.compute.manager [-] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Took 1.03 seconds to deallocate network for instance. [ 571.205293] env[61906]: DEBUG nova.compute.claims [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 571.205526] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.387690] env[61906]: DEBUG nova.network.neutron [req-488bb97b-d30c-43d3-a604-01cbfc9a683e req-7df1402e-2d84-4d80-8427-56fff6dd3658 service nova] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.420857] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.682597] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 7ed5b09c-5ba1-479c-96b4-75924df84e75 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 571.682879] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 571.684055] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 571.865166] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Acquiring lock "b4df3af4-f60d-4875-af62-d9c162e4a5c8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.865166] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Lock "b4df3af4-f60d-4875-af62-d9c162e4a5c8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.897268] env[61906]: DEBUG oslo_concurrency.lockutils [req-488bb97b-d30c-43d3-a604-01cbfc9a683e req-7df1402e-2d84-4d80-8427-56fff6dd3658 service nova] Releasing lock "refresh_cache-75ea63d7-6fc7-4a9f-b3df-05c6d12fe913" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 572.002118] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7018e89-69c5-412d-9104-74cde586fcf7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.013211] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0215440c-77a9-4ed7-968b-23c28684e881 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.045709] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf9f9e21-363c-44e1-919a-86cae305938c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.054891] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c61c8cc-ee87-4530-9e16-4394127d0d2b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.070913] env[61906]: DEBUG nova.compute.provider_tree [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 572.574502] env[61906]: DEBUG nova.scheduler.client.report [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 573.089020] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 573.089020] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.492s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.089020] env[61906]: DEBUG oslo_concurrency.lockutils [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.355s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.222557] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquiring lock "5b5e9435-ab0c-493a-8ddf-a480960a3818" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.222557] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "5b5e9435-ab0c-493a-8ddf-a480960a3818" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.251826] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquiring lock "cf331f9d-02ac-45d4-b938-0ed13c63e87b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.252018] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "cf331f9d-02ac-45d4-b938-0ed13c63e87b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.278409] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquiring lock "5eafc757-e09b-4751-ac50-3e918156832b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.278684] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "5eafc757-e09b-4751-ac50-3e918156832b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.970284] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83bfc303-40ec-48e5-800a-4807f8cd734a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.979932] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bed41bc3-216c-412c-8cb3-18b48c7d0b17 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.015168] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0b7a350-cd29-481c-b3ff-1c5e01770e5b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.023992] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-337e92c2-7c58-405b-9b3d-8fcae0224356 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.037525] env[61906]: DEBUG nova.compute.provider_tree [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 574.135548] env[61906]: DEBUG nova.compute.manager [req-659bf476-174b-49f3-9876-16857d030837 req-b9b91731-625e-4c7e-9467-3b69737e02e6 service nova] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Received event network-vif-deleted-0fb5cdf6-dbaf-4b77-aaf7-19f1195c7614 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 574.546256] env[61906]: DEBUG nova.scheduler.client.report [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 575.051231] env[61906]: DEBUG oslo_concurrency.lockutils [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.964s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.051871] env[61906]: ERROR nova.compute.manager [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fee6c75a-e766-4956-a3e2-1149ebf95e92, please check neutron logs for more information. [ 575.051871] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Traceback (most recent call last): [ 575.051871] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 575.051871] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] self.driver.spawn(context, instance, image_meta, [ 575.051871] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 575.051871] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] self._vmops.spawn(context, instance, image_meta, injected_files, [ 575.051871] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 575.051871] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] vm_ref = self.build_virtual_machine(instance, [ 575.051871] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 575.051871] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] vif_infos = vmwarevif.get_vif_info(self._session, [ 575.051871] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 575.052600] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] for vif in network_info: [ 575.052600] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 575.052600] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] return self._sync_wrapper(fn, *args, **kwargs) [ 575.052600] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 575.052600] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] self.wait() [ 575.052600] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 575.052600] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] self[:] = self._gt.wait() [ 575.052600] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 575.052600] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] return self._exit_event.wait() [ 575.052600] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 575.052600] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] result = hub.switch() [ 575.052600] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 575.052600] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] return self.greenlet.switch() [ 575.053273] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.053273] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] result = function(*args, **kwargs) [ 575.053273] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 575.053273] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] return func(*args, **kwargs) [ 575.053273] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 575.053273] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] raise e [ 575.053273] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.053273] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] nwinfo = self.network_api.allocate_for_instance( [ 575.053273] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 575.053273] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] created_port_ids = self._update_ports_for_instance( [ 575.053273] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 575.053273] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] with excutils.save_and_reraise_exception(): [ 575.053273] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.054243] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] self.force_reraise() [ 575.054243] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.054243] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] raise self.value [ 575.054243] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 575.054243] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] updated_port = self._update_port( [ 575.054243] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.054243] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] _ensure_no_port_binding_failure(port) [ 575.054243] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.054243] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] raise exception.PortBindingFailed(port_id=port['id']) [ 575.054243] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] nova.exception.PortBindingFailed: Binding failed for port fee6c75a-e766-4956-a3e2-1149ebf95e92, please check neutron logs for more information. [ 575.054243] env[61906]: ERROR nova.compute.manager [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] [ 575.054552] env[61906]: DEBUG nova.compute.utils [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Binding failed for port fee6c75a-e766-4956-a3e2-1149ebf95e92, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 575.054552] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.174s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.056692] env[61906]: INFO nova.compute.claims [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 575.059728] env[61906]: DEBUG nova.compute.manager [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Build of instance 402707da-f091-40fb-b5c7-3a2b1429ecce was re-scheduled: Binding failed for port fee6c75a-e766-4956-a3e2-1149ebf95e92, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 575.060138] env[61906]: DEBUG nova.compute.manager [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 575.060348] env[61906]: DEBUG oslo_concurrency.lockutils [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Acquiring lock "refresh_cache-402707da-f091-40fb-b5c7-3a2b1429ecce" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.060492] env[61906]: DEBUG oslo_concurrency.lockutils [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Acquired lock "refresh_cache-402707da-f091-40fb-b5c7-3a2b1429ecce" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.060645] env[61906]: DEBUG nova.network.neutron [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 575.589074] env[61906]: DEBUG nova.network.neutron [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 575.810224] env[61906]: DEBUG nova.network.neutron [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.314484] env[61906]: DEBUG oslo_concurrency.lockutils [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Releasing lock "refresh_cache-402707da-f091-40fb-b5c7-3a2b1429ecce" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.314768] env[61906]: DEBUG nova.compute.manager [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 576.314968] env[61906]: DEBUG nova.compute.manager [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 576.315190] env[61906]: DEBUG nova.network.neutron [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 576.359122] env[61906]: DEBUG nova.network.neutron [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 576.474019] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e666e5-b87a-4ab8-b542-4e8cb328b325 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.488097] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e2c29f2-5dad-406e-835c-fb852762cdf0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.530077] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9a61e42-1fe5-4fc1-b76f-fd5ffcae3d74 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.538642] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b1d95c5-04d3-4be1-b386-23cde5f8d70f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.555308] env[61906]: DEBUG nova.compute.provider_tree [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 576.861267] env[61906]: DEBUG nova.network.neutron [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.065569] env[61906]: DEBUG nova.scheduler.client.report [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 577.364522] env[61906]: INFO nova.compute.manager [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] [instance: 402707da-f091-40fb-b5c7-3a2b1429ecce] Took 1.05 seconds to deallocate network for instance. [ 577.568015] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.512s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.568015] env[61906]: DEBUG nova.compute.manager [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 577.572539] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.019s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.045970] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Acquiring lock "3595e6c6-10f3-46ba-85c4-90e25de78c1b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.046290] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Lock "3595e6c6-10f3-46ba-85c4-90e25de78c1b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.079680] env[61906]: DEBUG nova.compute.utils [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 578.087268] env[61906]: DEBUG nova.compute.manager [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 578.087268] env[61906]: DEBUG nova.network.neutron [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 578.215543] env[61906]: DEBUG nova.policy [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eb63f54f4e6549f29c8701643fc7b2fa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '079ccb1e022c4302a2a94f80a0507994', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 578.399971] env[61906]: INFO nova.scheduler.client.report [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Deleted allocations for instance 402707da-f091-40fb-b5c7-3a2b1429ecce [ 578.441082] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-030beda5-f613-4b4e-8e81-075c23bc79ee {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.450753] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7e4b146-27ab-4c31-ab5a-a9ca5a6666ea {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.484735] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-630b27dd-4a60-4c31-a6cf-848f68c503a1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.493568] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4afd219d-3246-4361-be9b-19b43829e522 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.509378] env[61906]: DEBUG nova.compute.provider_tree [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 578.588443] env[61906]: DEBUG nova.compute.manager [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 578.798092] env[61906]: DEBUG nova.network.neutron [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Successfully created port: 2b4a52ef-aea0-47fd-b004-fbe9aeb537b2 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 578.908246] env[61906]: DEBUG oslo_concurrency.lockutils [None req-287c0a84-8615-4c38-9594-e37be5358240 tempest-ServerMetadataTestJSON-194755405 tempest-ServerMetadataTestJSON-194755405-project-member] Lock "402707da-f091-40fb-b5c7-3a2b1429ecce" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.295s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.014742] env[61906]: DEBUG nova.scheduler.client.report [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 579.418604] env[61906]: DEBUG nova.compute.manager [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 579.524073] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.950s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.524073] env[61906]: ERROR nova.compute.manager [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 23b80c78-a492-49a8-a9e9-2d6e03a611fa, please check neutron logs for more information. [ 579.524073] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Traceback (most recent call last): [ 579.524073] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 579.524073] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] self.driver.spawn(context, instance, image_meta, [ 579.524073] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 579.524073] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 579.524073] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 579.524073] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] vm_ref = self.build_virtual_machine(instance, [ 579.524329] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 579.524329] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] vif_infos = vmwarevif.get_vif_info(self._session, [ 579.524329] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 579.524329] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] for vif in network_info: [ 579.524329] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 579.524329] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] return self._sync_wrapper(fn, *args, **kwargs) [ 579.524329] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 579.524329] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] self.wait() [ 579.524329] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 579.524329] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] self[:] = self._gt.wait() [ 579.524329] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 579.524329] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] return self._exit_event.wait() [ 579.524329] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 579.524649] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] result = hub.switch() [ 579.524649] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 579.524649] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] return self.greenlet.switch() [ 579.524649] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.524649] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] result = function(*args, **kwargs) [ 579.524649] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 579.524649] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] return func(*args, **kwargs) [ 579.524649] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 579.524649] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] raise e [ 579.524649] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.524649] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] nwinfo = self.network_api.allocate_for_instance( [ 579.524649] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 579.524649] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] created_port_ids = self._update_ports_for_instance( [ 579.524968] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 579.524968] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] with excutils.save_and_reraise_exception(): [ 579.524968] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.524968] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] self.force_reraise() [ 579.524968] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.524968] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] raise self.value [ 579.524968] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 579.524968] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] updated_port = self._update_port( [ 579.524968] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.524968] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] _ensure_no_port_binding_failure(port) [ 579.524968] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.524968] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] raise exception.PortBindingFailed(port_id=port['id']) [ 579.525293] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] nova.exception.PortBindingFailed: Binding failed for port 23b80c78-a492-49a8-a9e9-2d6e03a611fa, please check neutron logs for more information. [ 579.525293] env[61906]: ERROR nova.compute.manager [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] [ 579.525293] env[61906]: DEBUG nova.compute.utils [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Binding failed for port 23b80c78-a492-49a8-a9e9-2d6e03a611fa, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 579.525293] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.859s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.528145] env[61906]: INFO nova.compute.claims [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 579.531922] env[61906]: DEBUG nova.compute.manager [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Build of instance 1ac58495-75e5-4bce-8cb9-ecf18e0e436b was re-scheduled: Binding failed for port 23b80c78-a492-49a8-a9e9-2d6e03a611fa, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 579.532416] env[61906]: DEBUG nova.compute.manager [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 579.532655] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Acquiring lock "refresh_cache-1ac58495-75e5-4bce-8cb9-ecf18e0e436b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.532801] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Acquired lock "refresh_cache-1ac58495-75e5-4bce-8cb9-ecf18e0e436b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.532959] env[61906]: DEBUG nova.network.neutron [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 579.606324] env[61906]: DEBUG nova.compute.manager [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 579.632689] env[61906]: DEBUG nova.virt.hardware [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:20:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='798735576',id=21,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-421993576',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 579.632955] env[61906]: DEBUG nova.virt.hardware [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 579.633160] env[61906]: DEBUG nova.virt.hardware [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 579.633377] env[61906]: DEBUG nova.virt.hardware [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 579.633539] env[61906]: DEBUG nova.virt.hardware [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 579.633690] env[61906]: DEBUG nova.virt.hardware [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 579.633908] env[61906]: DEBUG nova.virt.hardware [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 579.634161] env[61906]: DEBUG nova.virt.hardware [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 579.634297] env[61906]: DEBUG nova.virt.hardware [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 579.635040] env[61906]: DEBUG nova.virt.hardware [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 579.635040] env[61906]: DEBUG nova.virt.hardware [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 579.635778] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff8432f4-576a-4fda-85bf-edea02dfd1bd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.652143] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-919d185a-7905-4a49-b218-e0fb357bf8d1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.951655] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.062041] env[61906]: DEBUG nova.network.neutron [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.123593] env[61906]: DEBUG nova.network.neutron [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.125303] env[61906]: DEBUG oslo_concurrency.lockutils [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Acquiring lock "32e2f482-9ceb-44bc-8933-115876374b1c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.125521] env[61906]: DEBUG oslo_concurrency.lockutils [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Lock "32e2f482-9ceb-44bc-8933-115876374b1c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.628165] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Releasing lock "refresh_cache-1ac58495-75e5-4bce-8cb9-ecf18e0e436b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.628742] env[61906]: DEBUG nova.compute.manager [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 580.628953] env[61906]: DEBUG nova.compute.manager [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 580.629286] env[61906]: DEBUG nova.network.neutron [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 580.659507] env[61906]: DEBUG nova.network.neutron [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.976468] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adf935bd-a9ae-4e30-8a54-0bb534a2f67c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.983907] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff23540-4763-4d2a-91cb-00066830ec39 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.020770] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-367075db-4f80-4087-ba24-79ec83d477e2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.029843] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Acquiring lock "b6f46e45-742b-4801-86f0-fd8eea3611da" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.030101] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Lock "b6f46e45-742b-4801-86f0-fd8eea3611da" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.031475] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c62ca0de-6e5a-4290-be69-1f32182ecbc2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.049172] env[61906]: DEBUG nova.compute.provider_tree [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 581.162961] env[61906]: DEBUG nova.network.neutron [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.294561] env[61906]: ERROR nova.compute.manager [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2b4a52ef-aea0-47fd-b004-fbe9aeb537b2, please check neutron logs for more information. [ 581.294561] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 581.294561] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 581.294561] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 581.294561] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.294561] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 581.294561] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.294561] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 581.294561] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.294561] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 581.294561] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.294561] env[61906]: ERROR nova.compute.manager raise self.value [ 581.294561] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.294561] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 581.294561] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.294561] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 581.295083] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.295083] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 581.295083] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2b4a52ef-aea0-47fd-b004-fbe9aeb537b2, please check neutron logs for more information. [ 581.295083] env[61906]: ERROR nova.compute.manager [ 581.295083] env[61906]: Traceback (most recent call last): [ 581.295083] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 581.295083] env[61906]: listener.cb(fileno) [ 581.295083] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 581.295083] env[61906]: result = function(*args, **kwargs) [ 581.295083] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 581.295083] env[61906]: return func(*args, **kwargs) [ 581.295083] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 581.295083] env[61906]: raise e [ 581.295083] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 581.295083] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 581.295083] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.295083] env[61906]: created_port_ids = self._update_ports_for_instance( [ 581.295083] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.295083] env[61906]: with excutils.save_and_reraise_exception(): [ 581.295083] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.295083] env[61906]: self.force_reraise() [ 581.295083] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.295083] env[61906]: raise self.value [ 581.295083] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.295083] env[61906]: updated_port = self._update_port( [ 581.295083] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.295083] env[61906]: _ensure_no_port_binding_failure(port) [ 581.295083] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.295083] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 581.295872] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 2b4a52ef-aea0-47fd-b004-fbe9aeb537b2, please check neutron logs for more information. [ 581.295872] env[61906]: Removing descriptor: 16 [ 581.295872] env[61906]: ERROR nova.compute.manager [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2b4a52ef-aea0-47fd-b004-fbe9aeb537b2, please check neutron logs for more information. [ 581.295872] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Traceback (most recent call last): [ 581.295872] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 581.295872] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] yield resources [ 581.295872] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 581.295872] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] self.driver.spawn(context, instance, image_meta, [ 581.295872] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 581.295872] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] self._vmops.spawn(context, instance, image_meta, injected_files, [ 581.295872] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 581.295872] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] vm_ref = self.build_virtual_machine(instance, [ 581.296262] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 581.296262] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] vif_infos = vmwarevif.get_vif_info(self._session, [ 581.296262] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 581.296262] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] for vif in network_info: [ 581.296262] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 581.296262] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] return self._sync_wrapper(fn, *args, **kwargs) [ 581.296262] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 581.296262] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] self.wait() [ 581.296262] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 581.296262] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] self[:] = self._gt.wait() [ 581.296262] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 581.296262] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] return self._exit_event.wait() [ 581.296262] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 581.296651] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] result = hub.switch() [ 581.296651] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 581.296651] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] return self.greenlet.switch() [ 581.296651] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 581.296651] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] result = function(*args, **kwargs) [ 581.296651] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 581.296651] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] return func(*args, **kwargs) [ 581.296651] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 581.296651] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] raise e [ 581.296651] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 581.296651] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] nwinfo = self.network_api.allocate_for_instance( [ 581.296651] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.296651] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] created_port_ids = self._update_ports_for_instance( [ 581.297063] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.297063] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] with excutils.save_and_reraise_exception(): [ 581.297063] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.297063] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] self.force_reraise() [ 581.297063] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.297063] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] raise self.value [ 581.297063] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.297063] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] updated_port = self._update_port( [ 581.297063] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.297063] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] _ensure_no_port_binding_failure(port) [ 581.297063] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.297063] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] raise exception.PortBindingFailed(port_id=port['id']) [ 581.297431] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] nova.exception.PortBindingFailed: Binding failed for port 2b4a52ef-aea0-47fd-b004-fbe9aeb537b2, please check neutron logs for more information. [ 581.297431] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] [ 581.297431] env[61906]: INFO nova.compute.manager [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Terminating instance [ 581.297834] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Acquiring lock "refresh_cache-75124494-73f0-4188-817b-1c7b4a85c8de" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 581.297989] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Acquired lock "refresh_cache-75124494-73f0-4188-817b-1c7b4a85c8de" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 581.298247] env[61906]: DEBUG nova.network.neutron [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 581.552073] env[61906]: DEBUG nova.scheduler.client.report [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 581.666329] env[61906]: INFO nova.compute.manager [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] [instance: 1ac58495-75e5-4bce-8cb9-ecf18e0e436b] Took 1.04 seconds to deallocate network for instance. [ 581.808898] env[61906]: DEBUG oslo_concurrency.lockutils [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Acquiring lock "0a1b97fe-42fb-47bd-bba2-aeb75c0710fb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.809314] env[61906]: DEBUG oslo_concurrency.lockutils [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Lock "0a1b97fe-42fb-47bd-bba2-aeb75c0710fb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.843148] env[61906]: DEBUG nova.network.neutron [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 581.936857] env[61906]: DEBUG nova.compute.manager [req-2551f79d-4cd2-42cc-8954-7962eda92dc8 req-7080e3ae-e7c6-44a7-aebe-a252fdf4d36e service nova] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Received event network-changed-2b4a52ef-aea0-47fd-b004-fbe9aeb537b2 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 581.937056] env[61906]: DEBUG nova.compute.manager [req-2551f79d-4cd2-42cc-8954-7962eda92dc8 req-7080e3ae-e7c6-44a7-aebe-a252fdf4d36e service nova] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Refreshing instance network info cache due to event network-changed-2b4a52ef-aea0-47fd-b004-fbe9aeb537b2. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 581.939219] env[61906]: DEBUG oslo_concurrency.lockutils [req-2551f79d-4cd2-42cc-8954-7962eda92dc8 req-7080e3ae-e7c6-44a7-aebe-a252fdf4d36e service nova] Acquiring lock "refresh_cache-75124494-73f0-4188-817b-1c7b4a85c8de" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 581.942618] env[61906]: DEBUG nova.network.neutron [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.057221] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.532s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.057734] env[61906]: DEBUG nova.compute.manager [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 582.061223] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.370s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.449724] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Releasing lock "refresh_cache-75124494-73f0-4188-817b-1c7b4a85c8de" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 582.449724] env[61906]: DEBUG nova.compute.manager [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 582.450030] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 582.453815] env[61906]: DEBUG oslo_concurrency.lockutils [req-2551f79d-4cd2-42cc-8954-7962eda92dc8 req-7080e3ae-e7c6-44a7-aebe-a252fdf4d36e service nova] Acquired lock "refresh_cache-75124494-73f0-4188-817b-1c7b4a85c8de" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.454110] env[61906]: DEBUG nova.network.neutron [req-2551f79d-4cd2-42cc-8954-7962eda92dc8 req-7080e3ae-e7c6-44a7-aebe-a252fdf4d36e service nova] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Refreshing network info cache for port 2b4a52ef-aea0-47fd-b004-fbe9aeb537b2 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 582.455197] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e40366a4-12cc-4869-aef9-c5ee74a035a8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.472798] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abe9d1a4-ffcb-4453-9446-91ea70dcd934 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.501641] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 75124494-73f0-4188-817b-1c7b4a85c8de could not be found. [ 582.501814] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 582.502470] env[61906]: INFO nova.compute.manager [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Took 0.05 seconds to destroy the instance on the hypervisor. [ 582.502470] env[61906]: DEBUG oslo.service.loopingcall [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 582.502470] env[61906]: DEBUG nova.compute.manager [-] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 582.502659] env[61906]: DEBUG nova.network.neutron [-] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 582.530499] env[61906]: DEBUG nova.network.neutron [-] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 582.565642] env[61906]: DEBUG nova.compute.utils [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 582.576986] env[61906]: DEBUG nova.compute.manager [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 582.581783] env[61906]: DEBUG nova.network.neutron [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 582.674092] env[61906]: DEBUG nova.policy [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b713431093b04b95826eb13be1cd4add', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c23791ef2b0e42eebee091731211b16b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 582.716474] env[61906]: INFO nova.scheduler.client.report [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Deleted allocations for instance 1ac58495-75e5-4bce-8cb9-ecf18e0e436b [ 582.981488] env[61906]: DEBUG nova.network.neutron [req-2551f79d-4cd2-42cc-8954-7962eda92dc8 req-7080e3ae-e7c6-44a7-aebe-a252fdf4d36e service nova] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.032474] env[61906]: DEBUG nova.network.neutron [-] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.040303] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39c8a172-e294-4848-8a1b-bc041f1c0ed8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.051768] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0c3ff6b-556c-4932-9526-3a4e210af3d8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.088363] env[61906]: DEBUG nova.compute.manager [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 583.091837] env[61906]: DEBUG nova.network.neutron [req-2551f79d-4cd2-42cc-8954-7962eda92dc8 req-7080e3ae-e7c6-44a7-aebe-a252fdf4d36e service nova] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.093320] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b150b7f9-d3fb-412d-b9ff-c91d80abd8fc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.103109] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faa55de4-087d-453f-b1bb-af46afd6d404 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.119208] env[61906]: DEBUG nova.compute.provider_tree [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 583.156032] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Acquiring lock "59887ef7-b2fc-486c-a368-cebde3313113" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.156032] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Lock "59887ef7-b2fc-486c-a368-cebde3313113" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.231461] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ef3091d0-51c4-45f6-99c8-81525a2ac0b0 tempest-ServerExternalEventsTest-1934197918 tempest-ServerExternalEventsTest-1934197918-project-member] Lock "1ac58495-75e5-4bce-8cb9-ecf18e0e436b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.346s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.271660] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Acquiring lock "ff6fa10c-c1cc-4e85-af43-c52f06e35974" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.271660] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Lock "ff6fa10c-c1cc-4e85-af43-c52f06e35974" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.536110] env[61906]: INFO nova.compute.manager [-] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Took 1.03 seconds to deallocate network for instance. [ 583.538725] env[61906]: DEBUG nova.compute.claims [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 583.538909] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.602159] env[61906]: DEBUG oslo_concurrency.lockutils [req-2551f79d-4cd2-42cc-8954-7962eda92dc8 req-7080e3ae-e7c6-44a7-aebe-a252fdf4d36e service nova] Releasing lock "refresh_cache-75124494-73f0-4188-817b-1c7b4a85c8de" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.613519] env[61906]: DEBUG nova.network.neutron [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Successfully created port: a196df79-d6c4-481d-b94f-7363593aabe3 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 583.622802] env[61906]: DEBUG nova.scheduler.client.report [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 583.734034] env[61906]: DEBUG nova.compute.manager [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 584.107522] env[61906]: DEBUG nova.compute.manager [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 584.131113] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.071s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.131747] env[61906]: ERROR nova.compute.manager [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d49dae9f-3dfd-40e8-8835-ec36ed215ed8, please check neutron logs for more information. [ 584.131747] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Traceback (most recent call last): [ 584.131747] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 584.131747] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] self.driver.spawn(context, instance, image_meta, [ 584.131747] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 584.131747] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 584.131747] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 584.131747] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] vm_ref = self.build_virtual_machine(instance, [ 584.131747] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 584.131747] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] vif_infos = vmwarevif.get_vif_info(self._session, [ 584.131747] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 584.132376] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] for vif in network_info: [ 584.132376] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 584.132376] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] return self._sync_wrapper(fn, *args, **kwargs) [ 584.132376] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 584.132376] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] self.wait() [ 584.132376] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 584.132376] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] self[:] = self._gt.wait() [ 584.132376] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 584.132376] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] return self._exit_event.wait() [ 584.132376] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 584.132376] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] result = hub.switch() [ 584.132376] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 584.132376] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] return self.greenlet.switch() [ 584.133884] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 584.133884] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] result = function(*args, **kwargs) [ 584.133884] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 584.133884] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] return func(*args, **kwargs) [ 584.133884] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 584.133884] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] raise e [ 584.133884] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 584.133884] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] nwinfo = self.network_api.allocate_for_instance( [ 584.133884] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 584.133884] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] created_port_ids = self._update_ports_for_instance( [ 584.133884] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 584.133884] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] with excutils.save_and_reraise_exception(): [ 584.133884] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.135421] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] self.force_reraise() [ 584.135421] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.135421] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] raise self.value [ 584.135421] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 584.135421] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] updated_port = self._update_port( [ 584.135421] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.135421] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] _ensure_no_port_binding_failure(port) [ 584.135421] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.135421] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] raise exception.PortBindingFailed(port_id=port['id']) [ 584.135421] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] nova.exception.PortBindingFailed: Binding failed for port d49dae9f-3dfd-40e8-8835-ec36ed215ed8, please check neutron logs for more information. [ 584.135421] env[61906]: ERROR nova.compute.manager [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] [ 584.137940] env[61906]: DEBUG nova.compute.utils [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Binding failed for port d49dae9f-3dfd-40e8-8835-ec36ed215ed8, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 584.137940] env[61906]: DEBUG oslo_concurrency.lockutils [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.100s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.137940] env[61906]: INFO nova.compute.claims [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 584.141915] env[61906]: DEBUG nova.compute.manager [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Build of instance 7044fd5f-a202-45a4-a4fb-73d20ef87cb5 was re-scheduled: Binding failed for port d49dae9f-3dfd-40e8-8835-ec36ed215ed8, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 584.141915] env[61906]: DEBUG nova.compute.manager [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 584.141915] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "refresh_cache-7044fd5f-a202-45a4-a4fb-73d20ef87cb5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.141915] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquired lock "refresh_cache-7044fd5f-a202-45a4-a4fb-73d20ef87cb5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.142092] env[61906]: DEBUG nova.network.neutron [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 584.145541] env[61906]: DEBUG nova.virt.hardware [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 584.145676] env[61906]: DEBUG nova.virt.hardware [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 584.145835] env[61906]: DEBUG nova.virt.hardware [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 584.146194] env[61906]: DEBUG nova.virt.hardware [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 584.146272] env[61906]: DEBUG nova.virt.hardware [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 584.146498] env[61906]: DEBUG nova.virt.hardware [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 584.146829] env[61906]: DEBUG nova.virt.hardware [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 584.147455] env[61906]: DEBUG nova.virt.hardware [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 584.147455] env[61906]: DEBUG nova.virt.hardware [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 584.147634] env[61906]: DEBUG nova.virt.hardware [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 584.147900] env[61906]: DEBUG nova.virt.hardware [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 584.148763] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-517da377-0a36-4168-8c06-f30e0b4a00d2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.161380] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d2a6e3-4ba9-4b52-8f62-febe5504d757 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.260390] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.517872] env[61906]: DEBUG nova.compute.manager [req-b70068b1-9238-440f-ba89-ab1e366fa1ff req-b07e766f-20ed-4e51-8e27-2b2faa20d04a service nova] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Received event network-vif-deleted-2b4a52ef-aea0-47fd-b004-fbe9aeb537b2 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 584.854354] env[61906]: DEBUG nova.network.neutron [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.047292] env[61906]: DEBUG nova.network.neutron [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.334147] env[61906]: ERROR nova.compute.manager [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a196df79-d6c4-481d-b94f-7363593aabe3, please check neutron logs for more information. [ 585.334147] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 585.334147] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.334147] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 585.334147] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 585.334147] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 585.334147] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 585.334147] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 585.334147] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.334147] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 585.334147] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.334147] env[61906]: ERROR nova.compute.manager raise self.value [ 585.334147] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 585.334147] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 585.334147] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.334147] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 585.334615] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.334615] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 585.334615] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a196df79-d6c4-481d-b94f-7363593aabe3, please check neutron logs for more information. [ 585.334615] env[61906]: ERROR nova.compute.manager [ 585.334615] env[61906]: Traceback (most recent call last): [ 585.334615] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 585.334615] env[61906]: listener.cb(fileno) [ 585.334615] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.334615] env[61906]: result = function(*args, **kwargs) [ 585.334615] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 585.334615] env[61906]: return func(*args, **kwargs) [ 585.334615] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 585.334615] env[61906]: raise e [ 585.334615] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.334615] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 585.334615] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 585.334615] env[61906]: created_port_ids = self._update_ports_for_instance( [ 585.334615] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 585.334615] env[61906]: with excutils.save_and_reraise_exception(): [ 585.334615] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.334615] env[61906]: self.force_reraise() [ 585.334615] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.334615] env[61906]: raise self.value [ 585.334615] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 585.334615] env[61906]: updated_port = self._update_port( [ 585.334615] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.334615] env[61906]: _ensure_no_port_binding_failure(port) [ 585.334615] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.334615] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 585.335370] env[61906]: nova.exception.PortBindingFailed: Binding failed for port a196df79-d6c4-481d-b94f-7363593aabe3, please check neutron logs for more information. [ 585.335370] env[61906]: Removing descriptor: 16 [ 585.335370] env[61906]: ERROR nova.compute.manager [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a196df79-d6c4-481d-b94f-7363593aabe3, please check neutron logs for more information. [ 585.335370] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Traceback (most recent call last): [ 585.335370] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 585.335370] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] yield resources [ 585.335370] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 585.335370] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] self.driver.spawn(context, instance, image_meta, [ 585.335370] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 585.335370] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 585.335370] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 585.335370] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] vm_ref = self.build_virtual_machine(instance, [ 585.335743] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 585.335743] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] vif_infos = vmwarevif.get_vif_info(self._session, [ 585.335743] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 585.335743] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] for vif in network_info: [ 585.335743] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 585.335743] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] return self._sync_wrapper(fn, *args, **kwargs) [ 585.335743] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 585.335743] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] self.wait() [ 585.335743] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 585.335743] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] self[:] = self._gt.wait() [ 585.335743] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 585.335743] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] return self._exit_event.wait() [ 585.335743] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 585.336171] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] result = hub.switch() [ 585.336171] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 585.336171] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] return self.greenlet.switch() [ 585.336171] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.336171] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] result = function(*args, **kwargs) [ 585.336171] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 585.336171] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] return func(*args, **kwargs) [ 585.336171] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 585.336171] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] raise e [ 585.336171] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.336171] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] nwinfo = self.network_api.allocate_for_instance( [ 585.336171] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 585.336171] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] created_port_ids = self._update_ports_for_instance( [ 585.336528] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 585.336528] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] with excutils.save_and_reraise_exception(): [ 585.336528] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.336528] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] self.force_reraise() [ 585.336528] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.336528] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] raise self.value [ 585.336528] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 585.336528] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] updated_port = self._update_port( [ 585.336528] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.336528] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] _ensure_no_port_binding_failure(port) [ 585.336528] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.336528] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] raise exception.PortBindingFailed(port_id=port['id']) [ 585.336889] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] nova.exception.PortBindingFailed: Binding failed for port a196df79-d6c4-481d-b94f-7363593aabe3, please check neutron logs for more information. [ 585.336889] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] [ 585.336889] env[61906]: INFO nova.compute.manager [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Terminating instance [ 585.338260] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Acquiring lock "refresh_cache-da9ac1bd-97c8-4c05-9364-8bd60be50ff2" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.338260] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Acquired lock "refresh_cache-da9ac1bd-97c8-4c05-9364-8bd60be50ff2" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.338260] env[61906]: DEBUG nova.network.neutron [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 585.552568] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Releasing lock "refresh_cache-7044fd5f-a202-45a4-a4fb-73d20ef87cb5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.552807] env[61906]: DEBUG nova.compute.manager [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 585.552991] env[61906]: DEBUG nova.compute.manager [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 585.553190] env[61906]: DEBUG nova.network.neutron [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 585.568575] env[61906]: DEBUG nova.network.neutron [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.607346] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0aed70d-5d5d-4254-8337-b832a654bf7b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.617364] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf88c12-e4f2-48e2-80f3-75193d81f923 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.654462] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffc48e21-f921-483b-abce-68c3491f408f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.664307] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-618c64e7-8d0a-4c64-83b2-2b70cc1cb3c8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.679267] env[61906]: DEBUG nova.compute.provider_tree [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 585.866577] env[61906]: DEBUG nova.network.neutron [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.925810] env[61906]: DEBUG nova.network.neutron [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.074196] env[61906]: DEBUG nova.network.neutron [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.184515] env[61906]: DEBUG nova.scheduler.client.report [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 586.432525] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Releasing lock "refresh_cache-da9ac1bd-97c8-4c05-9364-8bd60be50ff2" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.432836] env[61906]: DEBUG nova.compute.manager [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 586.432955] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 586.433277] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-409e1e9b-d8a1-4105-9932-6c428ab10308 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.448152] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d2a80a-dd6d-4e7f-9165-03bf0c7812d7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.484208] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance da9ac1bd-97c8-4c05-9364-8bd60be50ff2 could not be found. [ 586.484426] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 586.484617] env[61906]: INFO nova.compute.manager [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Took 0.05 seconds to destroy the instance on the hypervisor. [ 586.485262] env[61906]: DEBUG oslo.service.loopingcall [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 586.485487] env[61906]: DEBUG nova.compute.manager [-] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 586.486244] env[61906]: DEBUG nova.network.neutron [-] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 586.531126] env[61906]: DEBUG nova.network.neutron [-] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 586.580761] env[61906]: INFO nova.compute.manager [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 7044fd5f-a202-45a4-a4fb-73d20ef87cb5] Took 1.03 seconds to deallocate network for instance. [ 586.638161] env[61906]: DEBUG nova.compute.manager [req-e1075aed-6fd4-493c-b702-66f0d602598f req-129a9509-c52b-4aff-84a5-0c47804a4643 service nova] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Received event network-changed-a196df79-d6c4-481d-b94f-7363593aabe3 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 586.638161] env[61906]: DEBUG nova.compute.manager [req-e1075aed-6fd4-493c-b702-66f0d602598f req-129a9509-c52b-4aff-84a5-0c47804a4643 service nova] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Refreshing instance network info cache due to event network-changed-a196df79-d6c4-481d-b94f-7363593aabe3. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 586.638161] env[61906]: DEBUG oslo_concurrency.lockutils [req-e1075aed-6fd4-493c-b702-66f0d602598f req-129a9509-c52b-4aff-84a5-0c47804a4643 service nova] Acquiring lock "refresh_cache-da9ac1bd-97c8-4c05-9364-8bd60be50ff2" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.638161] env[61906]: DEBUG oslo_concurrency.lockutils [req-e1075aed-6fd4-493c-b702-66f0d602598f req-129a9509-c52b-4aff-84a5-0c47804a4643 service nova] Acquired lock "refresh_cache-da9ac1bd-97c8-4c05-9364-8bd60be50ff2" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.638161] env[61906]: DEBUG nova.network.neutron [req-e1075aed-6fd4-493c-b702-66f0d602598f req-129a9509-c52b-4aff-84a5-0c47804a4643 service nova] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Refreshing network info cache for port a196df79-d6c4-481d-b94f-7363593aabe3 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 586.692483] env[61906]: DEBUG oslo_concurrency.lockutils [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.558s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.693020] env[61906]: DEBUG nova.compute.manager [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 586.696753] env[61906]: DEBUG oslo_concurrency.lockutils [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.201s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.036154] env[61906]: DEBUG nova.network.neutron [-] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.195195] env[61906]: DEBUG nova.network.neutron [req-e1075aed-6fd4-493c-b702-66f0d602598f req-129a9509-c52b-4aff-84a5-0c47804a4643 service nova] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.201819] env[61906]: DEBUG nova.compute.utils [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 587.209873] env[61906]: DEBUG nova.compute.manager [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Not allocating networking since 'none' was specified. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 587.408630] env[61906]: DEBUG nova.network.neutron [req-e1075aed-6fd4-493c-b702-66f0d602598f req-129a9509-c52b-4aff-84a5-0c47804a4643 service nova] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.545217] env[61906]: INFO nova.compute.manager [-] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Took 1.06 seconds to deallocate network for instance. [ 587.549177] env[61906]: DEBUG nova.compute.claims [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 587.549177] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.615845] env[61906]: INFO nova.scheduler.client.report [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Deleted allocations for instance 7044fd5f-a202-45a4-a4fb-73d20ef87cb5 [ 587.689915] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de5bae23-272f-4355-ac35-6306a387d5cc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.699870] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93261eea-10f3-4486-8af0-f98f60239e69 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.740235] env[61906]: DEBUG nova.compute.manager [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 587.747213] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-381ee8ea-84f4-46e2-9255-0b8c7326745c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.756482] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab7d0e7-f1b2-4fe1-8ce3-280c7f0c2494 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.777132] env[61906]: DEBUG nova.compute.provider_tree [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 587.914444] env[61906]: DEBUG oslo_concurrency.lockutils [req-e1075aed-6fd4-493c-b702-66f0d602598f req-129a9509-c52b-4aff-84a5-0c47804a4643 service nova] Releasing lock "refresh_cache-da9ac1bd-97c8-4c05-9364-8bd60be50ff2" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.914444] env[61906]: DEBUG nova.compute.manager [req-e1075aed-6fd4-493c-b702-66f0d602598f req-129a9509-c52b-4aff-84a5-0c47804a4643 service nova] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Received event network-vif-deleted-a196df79-d6c4-481d-b94f-7363593aabe3 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 588.025940] env[61906]: DEBUG oslo_concurrency.lockutils [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Acquiring lock "410b0308-1dca-4d11-91ab-f6d89bd565e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.026206] env[61906]: DEBUG oslo_concurrency.lockutils [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Lock "410b0308-1dca-4d11-91ab-f6d89bd565e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.125585] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f3d5fe8-a7ee-41c0-b9aa-4ec7ec36601f tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "7044fd5f-a202-45a4-a4fb-73d20ef87cb5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.511s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.284021] env[61906]: DEBUG nova.scheduler.client.report [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 588.629088] env[61906]: DEBUG nova.compute.manager [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 588.758966] env[61906]: DEBUG nova.compute.manager [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 588.788439] env[61906]: DEBUG oslo_concurrency.lockutils [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.092s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.789275] env[61906]: ERROR nova.compute.manager [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9738c91a-842b-4cad-9068-a96785f62fd1, please check neutron logs for more information. [ 588.789275] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] Traceback (most recent call last): [ 588.789275] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 588.789275] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] self.driver.spawn(context, instance, image_meta, [ 588.789275] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 588.789275] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] self._vmops.spawn(context, instance, image_meta, injected_files, [ 588.789275] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 588.789275] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] vm_ref = self.build_virtual_machine(instance, [ 588.789275] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 588.789275] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] vif_infos = vmwarevif.get_vif_info(self._session, [ 588.789275] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 588.790503] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] for vif in network_info: [ 588.790503] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 588.790503] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] return self._sync_wrapper(fn, *args, **kwargs) [ 588.790503] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 588.790503] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] self.wait() [ 588.790503] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 588.790503] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] self[:] = self._gt.wait() [ 588.790503] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 588.790503] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] return self._exit_event.wait() [ 588.790503] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 588.790503] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] result = hub.switch() [ 588.790503] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 588.790503] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] return self.greenlet.switch() [ 588.790884] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.790884] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] result = function(*args, **kwargs) [ 588.790884] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 588.790884] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] return func(*args, **kwargs) [ 588.790884] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 588.790884] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] raise e [ 588.790884] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.790884] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] nwinfo = self.network_api.allocate_for_instance( [ 588.790884] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.790884] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] created_port_ids = self._update_ports_for_instance( [ 588.790884] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.790884] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] with excutils.save_and_reraise_exception(): [ 588.790884] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.791260] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] self.force_reraise() [ 588.791260] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.791260] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] raise self.value [ 588.791260] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.791260] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] updated_port = self._update_port( [ 588.791260] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.791260] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] _ensure_no_port_binding_failure(port) [ 588.791260] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.791260] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] raise exception.PortBindingFailed(port_id=port['id']) [ 588.791260] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] nova.exception.PortBindingFailed: Binding failed for port 9738c91a-842b-4cad-9068-a96785f62fd1, please check neutron logs for more information. [ 588.791260] env[61906]: ERROR nova.compute.manager [instance: 739af291-794c-491d-b579-2c50c5a14785] [ 588.791567] env[61906]: DEBUG nova.compute.utils [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Binding failed for port 9738c91a-842b-4cad-9068-a96785f62fd1, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 588.795036] env[61906]: DEBUG nova.virt.hardware [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 588.795256] env[61906]: DEBUG nova.virt.hardware [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 588.795415] env[61906]: DEBUG nova.virt.hardware [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 588.795592] env[61906]: DEBUG nova.virt.hardware [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 588.795727] env[61906]: DEBUG nova.virt.hardware [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 588.797195] env[61906]: DEBUG nova.virt.hardware [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 588.797195] env[61906]: DEBUG nova.virt.hardware [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 588.797810] env[61906]: DEBUG nova.virt.hardware [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 588.798033] env[61906]: DEBUG nova.virt.hardware [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 588.798202] env[61906]: DEBUG nova.virt.hardware [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 588.800066] env[61906]: DEBUG nova.virt.hardware [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 588.800066] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.743s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.803689] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2827dd25-6696-4cfe-82b2-48c64c7f896b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.808074] env[61906]: DEBUG nova.compute.manager [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Build of instance 739af291-794c-491d-b579-2c50c5a14785 was re-scheduled: Binding failed for port 9738c91a-842b-4cad-9068-a96785f62fd1, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 588.809026] env[61906]: DEBUG nova.compute.manager [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 588.809026] env[61906]: DEBUG oslo_concurrency.lockutils [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Acquiring lock "refresh_cache-739af291-794c-491d-b579-2c50c5a14785" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.809026] env[61906]: DEBUG oslo_concurrency.lockutils [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Acquired lock "refresh_cache-739af291-794c-491d-b579-2c50c5a14785" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.809026] env[61906]: DEBUG nova.network.neutron [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 588.816589] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a14ac7fe-a7d0-4840-8185-c243e2d83f7a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.837672] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Instance VIF info [] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 588.852199] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 588.852199] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4cfd5974-dab4-4eb5-a4ee-c1a064379bce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.865672] env[61906]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 588.865852] env[61906]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61906) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 588.866500] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Folder already exists: OpenStack. Parent ref: group-v4. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 588.866687] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Creating folder: Project (bc8d018a57644b329d7efbf10d077206). Parent ref: group-v284713. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 588.866910] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eca4f242-6016-4075-b573-3e5030324d2f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.879847] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Created folder: Project (bc8d018a57644b329d7efbf10d077206) in parent group-v284713. [ 588.880047] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Creating folder: Instances. Parent ref: group-v284718. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 588.880363] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b850ab46-2a3c-4112-9575-3cfbc04d3bbb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.893818] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Created folder: Instances in parent group-v284718. [ 588.894110] env[61906]: DEBUG oslo.service.loopingcall [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 588.894207] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 588.894427] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9525eb2c-a0c5-47e1-acea-9735b753d9d2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.913947] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 588.913947] env[61906]: value = "task-1333070" [ 588.913947] env[61906]: _type = "Task" [ 588.913947] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.922700] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333070, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 589.159602] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.372023] env[61906]: DEBUG nova.network.neutron [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.426878] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333070, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 589.525353] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Acquiring lock "2e8d82fd-5f70-408e-9d3b-74f3d81232d3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.525781] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Lock "2e8d82fd-5f70-408e-9d3b-74f3d81232d3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.790279] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-149b0096-40c1-41e7-bd73-c198707c49cd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.802202] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0283c917-9146-4867-b620-fd1866e4486c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.841967] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ab77f8-2a69-44d2-a3b5-d7c2fa825c4b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.846927] env[61906]: DEBUG nova.network.neutron [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.851898] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6100542b-ced5-45e5-83ab-a4fb75814351 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.869443] env[61906]: DEBUG nova.compute.provider_tree [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 589.926869] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333070, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 590.354996] env[61906]: DEBUG oslo_concurrency.lockutils [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Releasing lock "refresh_cache-739af291-794c-491d-b579-2c50c5a14785" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.354996] env[61906]: DEBUG nova.compute.manager [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 590.354996] env[61906]: DEBUG nova.compute.manager [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 590.354996] env[61906]: DEBUG nova.network.neutron [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 590.373310] env[61906]: DEBUG nova.scheduler.client.report [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 590.429115] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333070, 'name': CreateVM_Task, 'duration_secs': 1.363196} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 590.429944] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 590.431231] env[61906]: DEBUG oslo_vmware.service [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae75878b-a872-4ced-825c-a469134fc0b2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.437017] env[61906]: DEBUG nova.network.neutron [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.443512] env[61906]: DEBUG oslo_concurrency.lockutils [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.443751] env[61906]: DEBUG oslo_concurrency.lockutils [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.444582] env[61906]: DEBUG oslo_concurrency.lockutils [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 590.444897] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61cde299-35cc-4e42-b8a4-8238bf04c0b8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.451900] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Waiting for the task: (returnval){ [ 590.451900] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]527912b4-e1a0-93f1-5cc2-629ebecba184" [ 590.451900] env[61906]: _type = "Task" [ 590.451900] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 590.467103] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]527912b4-e1a0-93f1-5cc2-629ebecba184, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 590.880246] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.081s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.881196] env[61906]: ERROR nova.compute.manager [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3707cc06-d460-4cf9-b352-3c7dd250ea05, please check neutron logs for more information. [ 590.881196] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] Traceback (most recent call last): [ 590.881196] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 590.881196] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] self.driver.spawn(context, instance, image_meta, [ 590.881196] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 590.881196] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 590.881196] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 590.881196] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] vm_ref = self.build_virtual_machine(instance, [ 590.881196] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 590.881196] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] vif_infos = vmwarevif.get_vif_info(self._session, [ 590.881196] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 590.881605] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] for vif in network_info: [ 590.881605] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 590.881605] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] return self._sync_wrapper(fn, *args, **kwargs) [ 590.881605] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 590.881605] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] self.wait() [ 590.881605] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 590.881605] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] self[:] = self._gt.wait() [ 590.881605] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 590.881605] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] return self._exit_event.wait() [ 590.881605] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 590.881605] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] result = hub.switch() [ 590.881605] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 590.881605] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] return self.greenlet.switch() [ 590.882646] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 590.882646] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] result = function(*args, **kwargs) [ 590.882646] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 590.882646] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] return func(*args, **kwargs) [ 590.882646] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 590.882646] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] raise e [ 590.882646] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 590.882646] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] nwinfo = self.network_api.allocate_for_instance( [ 590.882646] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 590.882646] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] created_port_ids = self._update_ports_for_instance( [ 590.882646] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 590.882646] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] with excutils.save_and_reraise_exception(): [ 590.882646] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 590.883075] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] self.force_reraise() [ 590.883075] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 590.883075] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] raise self.value [ 590.883075] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 590.883075] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] updated_port = self._update_port( [ 590.883075] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 590.883075] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] _ensure_no_port_binding_failure(port) [ 590.883075] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 590.883075] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] raise exception.PortBindingFailed(port_id=port['id']) [ 590.883075] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] nova.exception.PortBindingFailed: Binding failed for port 3707cc06-d460-4cf9-b352-3c7dd250ea05, please check neutron logs for more information. [ 590.883075] env[61906]: ERROR nova.compute.manager [instance: bf480958-27a3-4450-b09c-cba96375712b] [ 590.883744] env[61906]: DEBUG nova.compute.utils [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Binding failed for port 3707cc06-d460-4cf9-b352-3c7dd250ea05, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 590.885252] env[61906]: DEBUG nova.compute.manager [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Build of instance bf480958-27a3-4450-b09c-cba96375712b was re-scheduled: Binding failed for port 3707cc06-d460-4cf9-b352-3c7dd250ea05, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 590.885722] env[61906]: DEBUG nova.compute.manager [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 590.886271] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Acquiring lock "refresh_cache-bf480958-27a3-4450-b09c-cba96375712b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.889346] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Acquired lock "refresh_cache-bf480958-27a3-4450-b09c-cba96375712b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.889595] env[61906]: DEBUG nova.network.neutron [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 590.890938] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.685s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.942216] env[61906]: DEBUG nova.network.neutron [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.968477] env[61906]: DEBUG oslo_concurrency.lockutils [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.969177] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 590.969523] env[61906]: DEBUG oslo_concurrency.lockutils [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.969749] env[61906]: DEBUG oslo_concurrency.lockutils [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.973247] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 590.973247] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d38a17e8-c6c2-403c-a9e6-8d8573312395 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.990307] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 590.990891] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 590.991701] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0e0a5e3-836d-4f96-ab83-4369071ee5a2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.999921] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-898c6b73-a718-4701-9c9c-ae9ee85d86d8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.006226] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Waiting for the task: (returnval){ [ 591.006226] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52b2fec5-f0ab-6812-dc8c-40bf6fd8e47c" [ 591.006226] env[61906]: _type = "Task" [ 591.006226] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 591.016754] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52b2fec5-f0ab-6812-dc8c-40bf6fd8e47c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 591.441321] env[61906]: DEBUG nova.network.neutron [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 591.448431] env[61906]: INFO nova.compute.manager [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] [instance: 739af291-794c-491d-b579-2c50c5a14785] Took 1.09 seconds to deallocate network for instance. [ 591.527913] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Preparing fetch location {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 591.528175] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Creating directory with path [datastore2] vmware_temp/d3e4ea56-2722-4c0f-ad55-8a5fe2b9a210/a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 591.529655] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4fa24ef1-dd16-41dd-85f9-83e75126e2d2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.570227] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Created directory with path [datastore2] vmware_temp/d3e4ea56-2722-4c0f-ad55-8a5fe2b9a210/a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 591.570227] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Fetch image to [datastore2] vmware_temp/d3e4ea56-2722-4c0f-ad55-8a5fe2b9a210/a9502563-1f96-46a5-b87e-12c0e9375e6e/tmp-sparse.vmdk {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 591.570227] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Downloading image file data a9502563-1f96-46a5-b87e-12c0e9375e6e to [datastore2] vmware_temp/d3e4ea56-2722-4c0f-ad55-8a5fe2b9a210/a9502563-1f96-46a5-b87e-12c0e9375e6e/tmp-sparse.vmdk on the data store datastore2 {{(pid=61906) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 591.571524] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1395f0f9-a78d-4369-af48-95752f2ad6de {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.583478] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b8af48-751e-4a42-9bdb-457c050bc56b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.596306] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfa40c21-27db-4d75-83f5-d28599b1cead {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.637433] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d363475-62e4-4247-a489-be507241c432 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.645463] env[61906]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-4f15c33e-9ce5-43d0-a789-2e82739a7a32 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.745458] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Downloading image file data a9502563-1f96-46a5-b87e-12c0e9375e6e to the data store datastore2 {{(pid=61906) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 591.820326] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "fd30cb23-f30f-4e20-ae67-36956b78e6ac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.820586] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "fd30cb23-f30f-4e20-ae67-36956b78e6ac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.834646] env[61906]: DEBUG oslo_vmware.rw_handles [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/d3e4ea56-2722-4c0f-ad55-8a5fe2b9a210/a9502563-1f96-46a5-b87e-12c0e9375e6e/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61906) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 591.895693] env[61906]: DEBUG nova.network.neutron [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.019052] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c98c8f2a-a52d-4bb9-94d3-9392eb51dda3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.029462] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41d8c58e-b6e0-4b4e-b7ba-f1b2bcbbb6e5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.066284] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-760453a7-465f-4853-818a-a1c95074340e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.075094] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70537c1c-1772-45f3-b9c5-51d28760561e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.089183] env[61906]: DEBUG nova.compute.provider_tree [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 592.402687] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Releasing lock "refresh_cache-bf480958-27a3-4450-b09c-cba96375712b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.402687] env[61906]: DEBUG nova.compute.manager [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 592.402687] env[61906]: DEBUG nova.compute.manager [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 592.402955] env[61906]: DEBUG nova.network.neutron [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 592.443176] env[61906]: DEBUG nova.network.neutron [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.493226] env[61906]: INFO nova.scheduler.client.report [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Deleted allocations for instance 739af291-794c-491d-b579-2c50c5a14785 [ 592.595945] env[61906]: DEBUG nova.scheduler.client.report [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 592.629054] env[61906]: DEBUG oslo_vmware.rw_handles [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Completed reading data from the image iterator. {{(pid=61906) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 592.629281] env[61906]: DEBUG oslo_vmware.rw_handles [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/d3e4ea56-2722-4c0f-ad55-8a5fe2b9a210/a9502563-1f96-46a5-b87e-12c0e9375e6e/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61906) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 592.772557] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Downloaded image file data a9502563-1f96-46a5-b87e-12c0e9375e6e to vmware_temp/d3e4ea56-2722-4c0f-ad55-8a5fe2b9a210/a9502563-1f96-46a5-b87e-12c0e9375e6e/tmp-sparse.vmdk on the data store datastore2 {{(pid=61906) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 592.774034] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Caching image {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 592.774500] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Copying Virtual Disk [datastore2] vmware_temp/d3e4ea56-2722-4c0f-ad55-8a5fe2b9a210/a9502563-1f96-46a5-b87e-12c0e9375e6e/tmp-sparse.vmdk to [datastore2] vmware_temp/d3e4ea56-2722-4c0f-ad55-8a5fe2b9a210/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 592.774500] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-88fea6d4-7bcf-424a-82a0-67cf8f2dd0ea {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.783129] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Waiting for the task: (returnval){ [ 592.783129] env[61906]: value = "task-1333071" [ 592.783129] env[61906]: _type = "Task" [ 592.783129] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.796568] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333071, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.948703] env[61906]: DEBUG nova.network.neutron [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.002723] env[61906]: DEBUG oslo_concurrency.lockutils [None req-27b9af0b-8f7a-480f-b58a-936177429607 tempest-ImagesOneServerNegativeTestJSON-396634808 tempest-ImagesOneServerNegativeTestJSON-396634808-project-member] Lock "739af291-794c-491d-b579-2c50c5a14785" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.730s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.103679] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.210s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.103679] env[61906]: ERROR nova.compute.manager [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0fb5cdf6-dbaf-4b77-aaf7-19f1195c7614, please check neutron logs for more information. [ 593.103679] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Traceback (most recent call last): [ 593.103679] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 593.103679] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] self.driver.spawn(context, instance, image_meta, [ 593.103679] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 593.103679] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] self._vmops.spawn(context, instance, image_meta, injected_files, [ 593.103679] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 593.103679] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] vm_ref = self.build_virtual_machine(instance, [ 593.104416] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 593.104416] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] vif_infos = vmwarevif.get_vif_info(self._session, [ 593.104416] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 593.104416] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] for vif in network_info: [ 593.104416] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 593.104416] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] return self._sync_wrapper(fn, *args, **kwargs) [ 593.104416] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 593.104416] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] self.wait() [ 593.104416] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 593.104416] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] self[:] = self._gt.wait() [ 593.104416] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 593.104416] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] return self._exit_event.wait() [ 593.104416] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 593.104805] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] result = hub.switch() [ 593.104805] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 593.104805] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] return self.greenlet.switch() [ 593.104805] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 593.104805] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] result = function(*args, **kwargs) [ 593.104805] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 593.104805] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] return func(*args, **kwargs) [ 593.104805] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 593.104805] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] raise e [ 593.104805] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.104805] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] nwinfo = self.network_api.allocate_for_instance( [ 593.104805] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 593.104805] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] created_port_ids = self._update_ports_for_instance( [ 593.105196] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 593.105196] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] with excutils.save_and_reraise_exception(): [ 593.105196] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.105196] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] self.force_reraise() [ 593.105196] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.105196] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] raise self.value [ 593.105196] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 593.105196] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] updated_port = self._update_port( [ 593.105196] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.105196] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] _ensure_no_port_binding_failure(port) [ 593.105196] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.105196] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] raise exception.PortBindingFailed(port_id=port['id']) [ 593.105609] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] nova.exception.PortBindingFailed: Binding failed for port 0fb5cdf6-dbaf-4b77-aaf7-19f1195c7614, please check neutron logs for more information. [ 593.105609] env[61906]: ERROR nova.compute.manager [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] [ 593.105609] env[61906]: DEBUG nova.compute.utils [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Binding failed for port 0fb5cdf6-dbaf-4b77-aaf7-19f1195c7614, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 593.108039] env[61906]: DEBUG nova.compute.manager [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Build of instance 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913 was re-scheduled: Binding failed for port 0fb5cdf6-dbaf-4b77-aaf7-19f1195c7614, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 593.108039] env[61906]: DEBUG nova.compute.manager [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 593.108355] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Acquiring lock "refresh_cache-75ea63d7-6fc7-4a9f-b3df-05c6d12fe913" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.108355] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Acquired lock "refresh_cache-75ea63d7-6fc7-4a9f-b3df-05c6d12fe913" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.108428] env[61906]: DEBUG nova.network.neutron [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 593.113095] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.689s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.113095] env[61906]: INFO nova.compute.claims [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 593.295079] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333071, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.452402] env[61906]: INFO nova.compute.manager [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] [instance: bf480958-27a3-4450-b09c-cba96375712b] Took 1.05 seconds to deallocate network for instance. [ 593.505869] env[61906]: DEBUG nova.compute.manager [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 593.665740] env[61906]: DEBUG nova.network.neutron [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.798665] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333071, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.726507} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.799114] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Copied Virtual Disk [datastore2] vmware_temp/d3e4ea56-2722-4c0f-ad55-8a5fe2b9a210/a9502563-1f96-46a5-b87e-12c0e9375e6e/tmp-sparse.vmdk to [datastore2] vmware_temp/d3e4ea56-2722-4c0f-ad55-8a5fe2b9a210/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 593.799464] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Deleting the datastore file [datastore2] vmware_temp/d3e4ea56-2722-4c0f-ad55-8a5fe2b9a210/a9502563-1f96-46a5-b87e-12c0e9375e6e/tmp-sparse.vmdk {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 593.799818] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7ebad47c-6ed8-4e0f-aeea-ac44c331fea4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.808666] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Waiting for the task: (returnval){ [ 593.808666] env[61906]: value = "task-1333072" [ 593.808666] env[61906]: _type = "Task" [ 593.808666] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.819327] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333072, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.040910] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.051141] env[61906]: DEBUG nova.network.neutron [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.323845] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333072, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.022266} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.324567] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 594.325123] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Moving file from [datastore2] vmware_temp/d3e4ea56-2722-4c0f-ad55-8a5fe2b9a210/a9502563-1f96-46a5-b87e-12c0e9375e6e to [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e. {{(pid=61906) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 594.326059] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-4dbd3862-f966-4df8-bc90-39150fe1c21b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.334349] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Waiting for the task: (returnval){ [ 594.334349] env[61906]: value = "task-1333073" [ 594.334349] env[61906]: _type = "Task" [ 594.334349] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.349839] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333073, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.510895] env[61906]: INFO nova.scheduler.client.report [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Deleted allocations for instance bf480958-27a3-4450-b09c-cba96375712b [ 594.558160] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Releasing lock "refresh_cache-75ea63d7-6fc7-4a9f-b3df-05c6d12fe913" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.558160] env[61906]: DEBUG nova.compute.manager [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 594.558718] env[61906]: DEBUG nova.compute.manager [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 594.558718] env[61906]: DEBUG nova.network.neutron [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 594.630226] env[61906]: DEBUG nova.network.neutron [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.688625] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9730ac6-600a-4940-bd24-4b039486f0be {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.703128] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9092429f-8fdc-4f9b-9156-ca467de97a3c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.743989] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-babb05ab-e0aa-4df5-b619-9c438d4f87bd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.751800] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06bc547f-d862-4425-8b9b-a44e5b2e986c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.774636] env[61906]: DEBUG nova.compute.provider_tree [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.852605] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333073, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.024578} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.852864] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] File moved {{(pid=61906) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 594.854523] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Cleaning up location [datastore2] vmware_temp/d3e4ea56-2722-4c0f-ad55-8a5fe2b9a210 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 594.854765] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Deleting the datastore file [datastore2] vmware_temp/d3e4ea56-2722-4c0f-ad55-8a5fe2b9a210 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 594.855526] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ca16971a-fb3d-4475-bea2-4d4a95c4fe4b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.866185] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Waiting for the task: (returnval){ [ 594.866185] env[61906]: value = "task-1333074" [ 594.866185] env[61906]: _type = "Task" [ 594.866185] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.877146] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333074, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.023122] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74962b75-fa33-4089-b2a4-f387e19b9413 tempest-ServerTagsTestJSON-2081439474 tempest-ServerTagsTestJSON-2081439474-project-member] Lock "bf480958-27a3-4450-b09c-cba96375712b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.404s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.097328] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Acquiring lock "fe14efda-3b92-4045-8eef-0e5e92a9c538" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.097556] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Lock "fe14efda-3b92-4045-8eef-0e5e92a9c538" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.136275] env[61906]: DEBUG nova.network.neutron [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.279902] env[61906]: DEBUG nova.scheduler.client.report [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 595.376054] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333074, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.039301} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 595.376480] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 595.377320] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-504f3519-5139-4b12-b00a-5b175b416da4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.383693] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Waiting for the task: (returnval){ [ 595.383693] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52f44e0e-b14a-1256-4e50-3d21b4628ae6" [ 595.383693] env[61906]: _type = "Task" [ 595.383693] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.394534] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52f44e0e-b14a-1256-4e50-3d21b4628ae6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.527364] env[61906]: DEBUG nova.compute.manager [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 595.641567] env[61906]: INFO nova.compute.manager [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913] Took 1.08 seconds to deallocate network for instance. [ 595.791864] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.679s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.791864] env[61906]: DEBUG nova.compute.manager [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 595.793203] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.844s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.794688] env[61906]: INFO nova.compute.claims [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 595.898279] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52f44e0e-b14a-1256-4e50-3d21b4628ae6, 'name': SearchDatastore_Task, 'duration_secs': 0.010412} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 595.898546] env[61906]: DEBUG oslo_concurrency.lockutils [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.899825] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] eb4cac2e-85d8-46b6-a4d4-ff12ee70e435/eb4cac2e-85d8-46b6-a4d4-ff12ee70e435.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 595.899825] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1ab7dd7c-93a6-49b2-ace1-deebe6573af3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.905740] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Waiting for the task: (returnval){ [ 595.905740] env[61906]: value = "task-1333075" [ 595.905740] env[61906]: _type = "Task" [ 595.905740] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.914196] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333075, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.065283] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.319098] env[61906]: DEBUG nova.compute.utils [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 596.324830] env[61906]: DEBUG nova.compute.manager [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 596.325029] env[61906]: DEBUG nova.network.neutron [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 596.422073] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333075, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.513358] env[61906]: DEBUG nova.policy [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03420f7f48b741c3988537dc6a5065e0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c625484c094f4fb1a38ed92fb39c4174', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 596.724311] env[61906]: INFO nova.scheduler.client.report [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Deleted allocations for instance 75ea63d7-6fc7-4a9f-b3df-05c6d12fe913 [ 596.834309] env[61906]: DEBUG nova.compute.manager [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 596.918991] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333075, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.631278} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 596.921946] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] eb4cac2e-85d8-46b6-a4d4-ff12ee70e435/eb4cac2e-85d8-46b6-a4d4-ff12ee70e435.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 596.922209] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 596.922734] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1e909744-352a-4aba-9015-ee1717f7f360 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.930721] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Waiting for the task: (returnval){ [ 596.930721] env[61906]: value = "task-1333076" [ 596.930721] env[61906]: _type = "Task" [ 596.930721] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 596.941983] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333076, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 597.240332] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea71880c-e797-40af-a244-38b12caff6f9 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Lock "75ea63d7-6fc7-4a9f-b3df-05c6d12fe913" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.370s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.253148] env[61906]: DEBUG oslo_concurrency.lockutils [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Acquiring lock "9e665b58-69e7-4ab9-b109-ab27725e66cc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.253375] env[61906]: DEBUG oslo_concurrency.lockutils [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Lock "9e665b58-69e7-4ab9-b109-ab27725e66cc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.384214] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77966dc9-5a46-4af2-884b-b981d7e5b3b3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.399522] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef92b2e0-a276-47bd-b1a2-6ee02a2d079c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.439742] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bce0245-de9a-42d9-b6a4-ef93747b8788 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.457985] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333076, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060764} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 597.459741] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ce4bcc2-b3ef-4170-9703-9b51ebcbb53a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.463452] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 597.465035] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d439b24-f9fe-43f8-9a52-2efab0955bf6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.480242] env[61906]: DEBUG nova.compute.provider_tree [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.502118] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Reconfiguring VM instance instance-0000000a to attach disk [datastore2] eb4cac2e-85d8-46b6-a4d4-ff12ee70e435/eb4cac2e-85d8-46b6-a4d4-ff12ee70e435.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 597.502118] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b631c853-e71e-40c1-abbd-6a9e9f217119 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.524749] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Waiting for the task: (returnval){ [ 597.524749] env[61906]: value = "task-1333077" [ 597.524749] env[61906]: _type = "Task" [ 597.524749] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 597.536538] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333077, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 597.741592] env[61906]: DEBUG nova.compute.manager [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 597.855839] env[61906]: DEBUG nova.compute.manager [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 597.874109] env[61906]: DEBUG nova.network.neutron [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Successfully created port: e59486b6-bb9d-4623-858d-2956318d9053 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 597.891014] env[61906]: DEBUG nova.virt.hardware [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 597.891272] env[61906]: DEBUG nova.virt.hardware [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 597.891426] env[61906]: DEBUG nova.virt.hardware [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 597.892221] env[61906]: DEBUG nova.virt.hardware [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 597.892221] env[61906]: DEBUG nova.virt.hardware [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 597.892221] env[61906]: DEBUG nova.virt.hardware [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 597.893555] env[61906]: DEBUG nova.virt.hardware [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 597.893785] env[61906]: DEBUG nova.virt.hardware [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 597.893964] env[61906]: DEBUG nova.virt.hardware [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 597.894617] env[61906]: DEBUG nova.virt.hardware [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 597.894853] env[61906]: DEBUG nova.virt.hardware [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 597.895839] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac570a17-5b30-47ae-9e87-ecc77f3c5d68 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.907901] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8e1240f-f6e5-42a8-80a4-924b8502f604 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.003770] env[61906]: DEBUG nova.scheduler.client.report [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 598.034951] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333077, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.271592] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.514765] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.721s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.515762] env[61906]: DEBUG nova.compute.manager [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 598.518540] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.980s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.538545] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333077, 'name': ReconfigVM_Task, 'duration_secs': 0.75255} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 598.539090] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Reconfigured VM instance instance-0000000a to attach disk [datastore2] eb4cac2e-85d8-46b6-a4d4-ff12ee70e435/eb4cac2e-85d8-46b6-a4d4-ff12ee70e435.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 598.540795] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d5cf60bc-bb57-486f-8e6d-6c6ad82396b6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.548825] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Waiting for the task: (returnval){ [ 598.548825] env[61906]: value = "task-1333078" [ 598.548825] env[61906]: _type = "Task" [ 598.548825] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 598.565014] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333078, 'name': Rename_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.029888] env[61906]: DEBUG nova.compute.utils [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 599.034122] env[61906]: DEBUG nova.compute.manager [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 599.034122] env[61906]: DEBUG nova.network.neutron [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 599.067475] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333078, 'name': Rename_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.169934] env[61906]: DEBUG nova.policy [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b713431093b04b95826eb13be1cd4add', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c23791ef2b0e42eebee091731211b16b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 599.490172] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf1b6faf-30b3-4712-9d4a-bc76d6b48aa5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.499623] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b93bb2f9-ed78-4aeb-aa7d-6be384fb86ec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.533659] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e3fbdc2-5f60-41c8-8f77-4d5150d2498d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.537269] env[61906]: DEBUG nova.compute.manager [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 599.543268] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aed8c0b5-3494-481e-979d-f5b6bcbab6c0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.557530] env[61906]: DEBUG nova.compute.provider_tree [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.566755] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333078, 'name': Rename_Task, 'duration_secs': 0.799236} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.567096] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 599.567375] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e95049e-dbb3-4677-a3ef-3b9aa2e58e29 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.577181] env[61906]: DEBUG oslo_concurrency.lockutils [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Acquiring lock "8d2459c0-279c-4ea4-8cbc-6c69cfa35b27" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.577418] env[61906]: DEBUG oslo_concurrency.lockutils [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Lock "8d2459c0-279c-4ea4-8cbc-6c69cfa35b27" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.577738] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Waiting for the task: (returnval){ [ 599.577738] env[61906]: value = "task-1333079" [ 599.577738] env[61906]: _type = "Task" [ 599.577738] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.589329] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333079, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.006686] env[61906]: DEBUG nova.network.neutron [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Successfully created port: d43156d0-1126-4b2d-8862-1fbd2a120665 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 600.055910] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Acquiring lock "7b4a9ebc-47fb-4078-969a-07642284b916" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.056372] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Lock "7b4a9ebc-47fb-4078-969a-07642284b916" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.063576] env[61906]: DEBUG nova.scheduler.client.report [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 600.090058] env[61906]: DEBUG oslo_vmware.api [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333079, 'name': PowerOnVM_Task, 'duration_secs': 0.40994} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.090588] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 600.090588] env[61906]: INFO nova.compute.manager [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Took 11.33 seconds to spawn the instance on the hypervisor. [ 600.090824] env[61906]: DEBUG nova.compute.manager [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 600.091599] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b39a221-b40b-4457-83ae-9c5571a9269d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.553207] env[61906]: DEBUG nova.compute.manager [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 600.571386] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.053s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.572226] env[61906]: ERROR nova.compute.manager [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2b4a52ef-aea0-47fd-b004-fbe9aeb537b2, please check neutron logs for more information. [ 600.572226] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Traceback (most recent call last): [ 600.572226] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 600.572226] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] self.driver.spawn(context, instance, image_meta, [ 600.572226] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 600.572226] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] self._vmops.spawn(context, instance, image_meta, injected_files, [ 600.572226] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 600.572226] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] vm_ref = self.build_virtual_machine(instance, [ 600.572226] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 600.572226] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] vif_infos = vmwarevif.get_vif_info(self._session, [ 600.572226] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 600.572643] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] for vif in network_info: [ 600.572643] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 600.572643] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] return self._sync_wrapper(fn, *args, **kwargs) [ 600.572643] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 600.572643] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] self.wait() [ 600.572643] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 600.572643] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] self[:] = self._gt.wait() [ 600.572643] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 600.572643] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] return self._exit_event.wait() [ 600.572643] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 600.572643] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] result = hub.switch() [ 600.572643] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 600.572643] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] return self.greenlet.switch() [ 600.573018] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.573018] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] result = function(*args, **kwargs) [ 600.573018] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 600.573018] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] return func(*args, **kwargs) [ 600.573018] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 600.573018] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] raise e [ 600.573018] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.573018] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] nwinfo = self.network_api.allocate_for_instance( [ 600.573018] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 600.573018] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] created_port_ids = self._update_ports_for_instance( [ 600.573018] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 600.573018] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] with excutils.save_and_reraise_exception(): [ 600.573018] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.573386] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] self.force_reraise() [ 600.573386] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.573386] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] raise self.value [ 600.573386] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 600.573386] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] updated_port = self._update_port( [ 600.573386] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.573386] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] _ensure_no_port_binding_failure(port) [ 600.573386] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.573386] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] raise exception.PortBindingFailed(port_id=port['id']) [ 600.573386] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] nova.exception.PortBindingFailed: Binding failed for port 2b4a52ef-aea0-47fd-b004-fbe9aeb537b2, please check neutron logs for more information. [ 600.573386] env[61906]: ERROR nova.compute.manager [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] [ 600.573831] env[61906]: DEBUG nova.compute.utils [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Binding failed for port 2b4a52ef-aea0-47fd-b004-fbe9aeb537b2, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 600.576173] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.315s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.577724] env[61906]: INFO nova.compute.claims [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 600.582024] env[61906]: DEBUG nova.compute.manager [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Build of instance 75124494-73f0-4188-817b-1c7b4a85c8de was re-scheduled: Binding failed for port 2b4a52ef-aea0-47fd-b004-fbe9aeb537b2, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 600.582024] env[61906]: DEBUG nova.compute.manager [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 600.582024] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Acquiring lock "refresh_cache-75124494-73f0-4188-817b-1c7b4a85c8de" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.582024] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Acquired lock "refresh_cache-75124494-73f0-4188-817b-1c7b4a85c8de" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.582281] env[61906]: DEBUG nova.network.neutron [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 600.596601] env[61906]: DEBUG nova.virt.hardware [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 600.598789] env[61906]: DEBUG nova.virt.hardware [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 600.598928] env[61906]: DEBUG nova.virt.hardware [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 600.599325] env[61906]: DEBUG nova.virt.hardware [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 600.599830] env[61906]: DEBUG nova.virt.hardware [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 600.599830] env[61906]: DEBUG nova.virt.hardware [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 600.600071] env[61906]: DEBUG nova.virt.hardware [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 600.600337] env[61906]: DEBUG nova.virt.hardware [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 600.600564] env[61906]: DEBUG nova.virt.hardware [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 600.600890] env[61906]: DEBUG nova.virt.hardware [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 600.601188] env[61906]: DEBUG nova.virt.hardware [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 600.606922] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-482d67c8-0416-435d-bd84-a0d7b2070e9a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.621297] env[61906]: INFO nova.compute.manager [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Took 34.61 seconds to build instance. [ 600.621297] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-729b4891-4b1d-4d8f-a610-c60b7b82581c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.126255] env[61906]: DEBUG oslo_concurrency.lockutils [None req-57c2eaa2-eeaa-4298-80b3-be24dece4fa8 tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Lock "eb4cac2e-85d8-46b6-a4d4-ff12ee70e435" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.133s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.140695] env[61906]: DEBUG nova.network.neutron [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.334023] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "7c389914-47a6-4ed4-aad6-e8aeeefbed7c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.334023] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "7c389914-47a6-4ed4-aad6-e8aeeefbed7c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.381502] env[61906]: DEBUG nova.network.neutron [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.548574] env[61906]: ERROR nova.compute.manager [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e59486b6-bb9d-4623-858d-2956318d9053, please check neutron logs for more information. [ 601.548574] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 601.548574] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.548574] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 601.548574] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.548574] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 601.548574] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.548574] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 601.548574] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.548574] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 601.548574] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.548574] env[61906]: ERROR nova.compute.manager raise self.value [ 601.548574] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.548574] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 601.548574] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.548574] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 601.549382] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.549382] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 601.549382] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e59486b6-bb9d-4623-858d-2956318d9053, please check neutron logs for more information. [ 601.549382] env[61906]: ERROR nova.compute.manager [ 601.549382] env[61906]: Traceback (most recent call last): [ 601.549382] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 601.549382] env[61906]: listener.cb(fileno) [ 601.549382] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.549382] env[61906]: result = function(*args, **kwargs) [ 601.549382] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 601.549382] env[61906]: return func(*args, **kwargs) [ 601.549382] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.549382] env[61906]: raise e [ 601.549382] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.549382] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 601.549382] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.549382] env[61906]: created_port_ids = self._update_ports_for_instance( [ 601.549382] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.549382] env[61906]: with excutils.save_and_reraise_exception(): [ 601.549382] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.549382] env[61906]: self.force_reraise() [ 601.549382] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.549382] env[61906]: raise self.value [ 601.549382] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.549382] env[61906]: updated_port = self._update_port( [ 601.549382] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.549382] env[61906]: _ensure_no_port_binding_failure(port) [ 601.549382] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.549382] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 601.550343] env[61906]: nova.exception.PortBindingFailed: Binding failed for port e59486b6-bb9d-4623-858d-2956318d9053, please check neutron logs for more information. [ 601.550343] env[61906]: Removing descriptor: 16 [ 601.550343] env[61906]: ERROR nova.compute.manager [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e59486b6-bb9d-4623-858d-2956318d9053, please check neutron logs for more information. [ 601.550343] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] Traceback (most recent call last): [ 601.550343] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 601.550343] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] yield resources [ 601.550343] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 601.550343] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] self.driver.spawn(context, instance, image_meta, [ 601.550343] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 601.550343] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] self._vmops.spawn(context, instance, image_meta, injected_files, [ 601.550343] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 601.550343] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] vm_ref = self.build_virtual_machine(instance, [ 601.550801] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 601.550801] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] vif_infos = vmwarevif.get_vif_info(self._session, [ 601.550801] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 601.550801] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] for vif in network_info: [ 601.550801] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 601.550801] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] return self._sync_wrapper(fn, *args, **kwargs) [ 601.550801] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 601.550801] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] self.wait() [ 601.550801] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 601.550801] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] self[:] = self._gt.wait() [ 601.550801] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 601.550801] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] return self._exit_event.wait() [ 601.550801] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 601.551252] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] result = hub.switch() [ 601.551252] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 601.551252] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] return self.greenlet.switch() [ 601.551252] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.551252] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] result = function(*args, **kwargs) [ 601.551252] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 601.551252] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] return func(*args, **kwargs) [ 601.551252] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.551252] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] raise e [ 601.551252] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.551252] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] nwinfo = self.network_api.allocate_for_instance( [ 601.551252] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.551252] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] created_port_ids = self._update_ports_for_instance( [ 601.551806] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.551806] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] with excutils.save_and_reraise_exception(): [ 601.551806] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.551806] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] self.force_reraise() [ 601.551806] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.551806] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] raise self.value [ 601.551806] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.551806] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] updated_port = self._update_port( [ 601.551806] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.551806] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] _ensure_no_port_binding_failure(port) [ 601.551806] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.551806] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] raise exception.PortBindingFailed(port_id=port['id']) [ 601.552222] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] nova.exception.PortBindingFailed: Binding failed for port e59486b6-bb9d-4623-858d-2956318d9053, please check neutron logs for more information. [ 601.552222] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] [ 601.552222] env[61906]: INFO nova.compute.manager [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Terminating instance [ 601.552222] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Acquiring lock "refresh_cache-263da5ea-892b-4142-b487-96444abc97be" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.552371] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Acquired lock "refresh_cache-263da5ea-892b-4142-b487-96444abc97be" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.552475] env[61906]: DEBUG nova.network.neutron [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 601.629470] env[61906]: DEBUG nova.compute.manager [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 601.888830] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Releasing lock "refresh_cache-75124494-73f0-4188-817b-1c7b4a85c8de" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.891690] env[61906]: DEBUG nova.compute.manager [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 601.891690] env[61906]: DEBUG nova.compute.manager [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 601.891690] env[61906]: DEBUG nova.network.neutron [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 601.922442] env[61906]: DEBUG nova.network.neutron [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.989626] env[61906]: DEBUG nova.compute.manager [req-f8266946-2ce0-4148-a17f-ef7fdbc972ed req-bebc220c-bc39-40a8-a25f-0554e943e06a service nova] [instance: 263da5ea-892b-4142-b487-96444abc97be] Received event network-changed-e59486b6-bb9d-4623-858d-2956318d9053 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 601.989834] env[61906]: DEBUG nova.compute.manager [req-f8266946-2ce0-4148-a17f-ef7fdbc972ed req-bebc220c-bc39-40a8-a25f-0554e943e06a service nova] [instance: 263da5ea-892b-4142-b487-96444abc97be] Refreshing instance network info cache due to event network-changed-e59486b6-bb9d-4623-858d-2956318d9053. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 601.990061] env[61906]: DEBUG oslo_concurrency.lockutils [req-f8266946-2ce0-4148-a17f-ef7fdbc972ed req-bebc220c-bc39-40a8-a25f-0554e943e06a service nova] Acquiring lock "refresh_cache-263da5ea-892b-4142-b487-96444abc97be" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.088767] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2e908eb-d548-4ba7-96f7-476d747c3f01 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.097232] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0da7198-afa9-443a-a46a-f83e862c0758 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.135178] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27b15005-fc27-4829-ab28-fd51119df43c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.149025] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af621725-eebb-4632-965c-dc55d09019b8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.154410] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.164330] env[61906]: DEBUG nova.compute.provider_tree [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 602.175953] env[61906]: DEBUG nova.network.neutron [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.302404] env[61906]: ERROR nova.compute.manager [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d43156d0-1126-4b2d-8862-1fbd2a120665, please check neutron logs for more information. [ 602.302404] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 602.302404] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.302404] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 602.302404] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 602.302404] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 602.302404] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 602.302404] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 602.302404] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.302404] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 602.302404] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.302404] env[61906]: ERROR nova.compute.manager raise self.value [ 602.302404] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 602.302404] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 602.302404] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.302404] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 602.303208] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.303208] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 602.303208] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d43156d0-1126-4b2d-8862-1fbd2a120665, please check neutron logs for more information. [ 602.303208] env[61906]: ERROR nova.compute.manager [ 602.303208] env[61906]: Traceback (most recent call last): [ 602.303208] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 602.303208] env[61906]: listener.cb(fileno) [ 602.303208] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.303208] env[61906]: result = function(*args, **kwargs) [ 602.303208] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 602.303208] env[61906]: return func(*args, **kwargs) [ 602.303208] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.303208] env[61906]: raise e [ 602.303208] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.303208] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 602.303208] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 602.303208] env[61906]: created_port_ids = self._update_ports_for_instance( [ 602.303208] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 602.303208] env[61906]: with excutils.save_and_reraise_exception(): [ 602.303208] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.303208] env[61906]: self.force_reraise() [ 602.303208] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.303208] env[61906]: raise self.value [ 602.303208] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 602.303208] env[61906]: updated_port = self._update_port( [ 602.303208] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.303208] env[61906]: _ensure_no_port_binding_failure(port) [ 602.303208] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.303208] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 602.305407] env[61906]: nova.exception.PortBindingFailed: Binding failed for port d43156d0-1126-4b2d-8862-1fbd2a120665, please check neutron logs for more information. [ 602.305407] env[61906]: Removing descriptor: 18 [ 602.305407] env[61906]: ERROR nova.compute.manager [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d43156d0-1126-4b2d-8862-1fbd2a120665, please check neutron logs for more information. [ 602.305407] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Traceback (most recent call last): [ 602.305407] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 602.305407] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] yield resources [ 602.305407] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 602.305407] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] self.driver.spawn(context, instance, image_meta, [ 602.305407] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 602.305407] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] self._vmops.spawn(context, instance, image_meta, injected_files, [ 602.305407] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 602.305407] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] vm_ref = self.build_virtual_machine(instance, [ 602.305934] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 602.305934] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] vif_infos = vmwarevif.get_vif_info(self._session, [ 602.305934] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 602.305934] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] for vif in network_info: [ 602.305934] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 602.305934] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] return self._sync_wrapper(fn, *args, **kwargs) [ 602.305934] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 602.305934] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] self.wait() [ 602.305934] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 602.305934] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] self[:] = self._gt.wait() [ 602.305934] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 602.305934] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] return self._exit_event.wait() [ 602.305934] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 602.306325] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] result = hub.switch() [ 602.306325] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 602.306325] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] return self.greenlet.switch() [ 602.306325] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.306325] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] result = function(*args, **kwargs) [ 602.306325] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 602.306325] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] return func(*args, **kwargs) [ 602.306325] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.306325] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] raise e [ 602.306325] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.306325] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] nwinfo = self.network_api.allocate_for_instance( [ 602.306325] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 602.306325] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] created_port_ids = self._update_ports_for_instance( [ 602.306693] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 602.306693] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] with excutils.save_and_reraise_exception(): [ 602.306693] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.306693] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] self.force_reraise() [ 602.306693] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.306693] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] raise self.value [ 602.306693] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 602.306693] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] updated_port = self._update_port( [ 602.306693] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.306693] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] _ensure_no_port_binding_failure(port) [ 602.306693] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.306693] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] raise exception.PortBindingFailed(port_id=port['id']) [ 602.307446] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] nova.exception.PortBindingFailed: Binding failed for port d43156d0-1126-4b2d-8862-1fbd2a120665, please check neutron logs for more information. [ 602.307446] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] [ 602.307446] env[61906]: INFO nova.compute.manager [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Terminating instance [ 602.307446] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Acquiring lock "refresh_cache-1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.307446] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Acquired lock "refresh_cache-1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.307446] env[61906]: DEBUG nova.network.neutron [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 602.366109] env[61906]: DEBUG nova.network.neutron [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.426767] env[61906]: DEBUG nova.network.neutron [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.521300] env[61906]: DEBUG oslo_concurrency.lockutils [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Acquiring lock "c15bc75b-87dc-4602-81ea-dbebce1b7763" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.521612] env[61906]: DEBUG oslo_concurrency.lockutils [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Lock "c15bc75b-87dc-4602-81ea-dbebce1b7763" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.672053] env[61906]: DEBUG nova.scheduler.client.report [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 602.838172] env[61906]: DEBUG nova.network.neutron [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.872763] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Releasing lock "refresh_cache-263da5ea-892b-4142-b487-96444abc97be" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.873797] env[61906]: DEBUG nova.compute.manager [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 602.874016] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 602.874352] env[61906]: DEBUG oslo_concurrency.lockutils [req-f8266946-2ce0-4148-a17f-ef7fdbc972ed req-bebc220c-bc39-40a8-a25f-0554e943e06a service nova] Acquired lock "refresh_cache-263da5ea-892b-4142-b487-96444abc97be" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.874525] env[61906]: DEBUG nova.network.neutron [req-f8266946-2ce0-4148-a17f-ef7fdbc972ed req-bebc220c-bc39-40a8-a25f-0554e943e06a service nova] [instance: 263da5ea-892b-4142-b487-96444abc97be] Refreshing network info cache for port e59486b6-bb9d-4623-858d-2956318d9053 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 602.875910] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b12d6831-5f1d-4390-8ced-cfd64914c846 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.887161] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f7fe6a1-4c4d-4d5c-97ca-f20544ce1113 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.915624] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 263da5ea-892b-4142-b487-96444abc97be could not be found. [ 602.915624] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 602.915624] env[61906]: INFO nova.compute.manager [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Took 0.04 seconds to destroy the instance on the hypervisor. [ 602.915624] env[61906]: DEBUG oslo.service.loopingcall [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 602.915624] env[61906]: DEBUG nova.compute.manager [-] [instance: 263da5ea-892b-4142-b487-96444abc97be] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 602.915624] env[61906]: DEBUG nova.network.neutron [-] [instance: 263da5ea-892b-4142-b487-96444abc97be] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 602.930176] env[61906]: INFO nova.compute.manager [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] [instance: 75124494-73f0-4188-817b-1c7b4a85c8de] Took 1.04 seconds to deallocate network for instance. [ 602.941276] env[61906]: DEBUG nova.network.neutron [-] [instance: 263da5ea-892b-4142-b487-96444abc97be] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.963155] env[61906]: DEBUG nova.network.neutron [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.152913] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Acquiring lock "c9cb7335-8977-4ddf-9d0e-ff2166abf8e7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.153154] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Lock "c9cb7335-8977-4ddf-9d0e-ff2166abf8e7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.177072] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.602s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.178091] env[61906]: DEBUG nova.compute.manager [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 603.181980] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.633s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.403681] env[61906]: DEBUG nova.compute.manager [req-c579c841-c461-42c9-8142-f75485f8b868 req-f51241c5-a773-42d2-8496-d56e24a665dc service nova] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Received event network-changed-d43156d0-1126-4b2d-8862-1fbd2a120665 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 603.403925] env[61906]: DEBUG nova.compute.manager [req-c579c841-c461-42c9-8142-f75485f8b868 req-f51241c5-a773-42d2-8496-d56e24a665dc service nova] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Refreshing instance network info cache due to event network-changed-d43156d0-1126-4b2d-8862-1fbd2a120665. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 603.403964] env[61906]: DEBUG oslo_concurrency.lockutils [req-c579c841-c461-42c9-8142-f75485f8b868 req-f51241c5-a773-42d2-8496-d56e24a665dc service nova] Acquiring lock "refresh_cache-1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.408324] env[61906]: DEBUG nova.network.neutron [req-f8266946-2ce0-4148-a17f-ef7fdbc972ed req-bebc220c-bc39-40a8-a25f-0554e943e06a service nova] [instance: 263da5ea-892b-4142-b487-96444abc97be] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.444062] env[61906]: DEBUG nova.network.neutron [-] [instance: 263da5ea-892b-4142-b487-96444abc97be] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.463727] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Releasing lock "refresh_cache-1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.463953] env[61906]: DEBUG nova.compute.manager [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 603.464316] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 603.464758] env[61906]: DEBUG oslo_concurrency.lockutils [req-c579c841-c461-42c9-8142-f75485f8b868 req-f51241c5-a773-42d2-8496-d56e24a665dc service nova] Acquired lock "refresh_cache-1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.465769] env[61906]: DEBUG nova.network.neutron [req-c579c841-c461-42c9-8142-f75485f8b868 req-f51241c5-a773-42d2-8496-d56e24a665dc service nova] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Refreshing network info cache for port d43156d0-1126-4b2d-8862-1fbd2a120665 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 603.465928] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-591ec20d-5492-42be-b8c4-958d76befe76 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.477897] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd109c5-751e-41e6-abb4-b16c749ca98b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.503166] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121 could not be found. [ 603.503302] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 603.503448] env[61906]: INFO nova.compute.manager [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Took 0.04 seconds to destroy the instance on the hypervisor. [ 603.503690] env[61906]: DEBUG oslo.service.loopingcall [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 603.503894] env[61906]: DEBUG nova.compute.manager [-] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 603.504417] env[61906]: DEBUG nova.network.neutron [-] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 603.540169] env[61906]: DEBUG oslo_concurrency.lockutils [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Acquiring lock "c4b41c42-e01f-464a-9423-1c0fcca75586" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.540447] env[61906]: DEBUG oslo_concurrency.lockutils [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Lock "c4b41c42-e01f-464a-9423-1c0fcca75586" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.541200] env[61906]: DEBUG nova.network.neutron [-] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.576594] env[61906]: DEBUG nova.network.neutron [req-f8266946-2ce0-4148-a17f-ef7fdbc972ed req-bebc220c-bc39-40a8-a25f-0554e943e06a service nova] [instance: 263da5ea-892b-4142-b487-96444abc97be] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.649448] env[61906]: DEBUG nova.compute.manager [None req-36379539-488e-4925-bd54-f7620f880e4c tempest-ServerDiagnosticsV248Test-447830524 tempest-ServerDiagnosticsV248Test-447830524-project-admin] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 603.651574] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77ecd986-2f5f-4297-8e8e-c3dfbe082987 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.659232] env[61906]: INFO nova.compute.manager [None req-36379539-488e-4925-bd54-f7620f880e4c tempest-ServerDiagnosticsV248Test-447830524 tempest-ServerDiagnosticsV248Test-447830524-project-admin] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Retrieving diagnostics [ 603.660419] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33956aa6-c652-4779-a875-6d964fdd86f8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.702884] env[61906]: DEBUG nova.compute.utils [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 603.709110] env[61906]: DEBUG nova.compute.manager [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Not allocating networking since 'none' was specified. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 603.712313] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "adcd5674-ce1e-4b73-85bd-4344b5633c94" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.712526] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "adcd5674-ce1e-4b73-85bd-4344b5633c94" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.729471] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Acquiring lock "bf056a94-9e9b-4296-b621-7e8a7338330b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.729763] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Lock "bf056a94-9e9b-4296-b621-7e8a7338330b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.946746] env[61906]: INFO nova.compute.manager [-] [instance: 263da5ea-892b-4142-b487-96444abc97be] Took 1.03 seconds to deallocate network for instance. [ 603.950173] env[61906]: DEBUG nova.compute.claims [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 603.950173] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.965442] env[61906]: INFO nova.scheduler.client.report [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Deleted allocations for instance 75124494-73f0-4188-817b-1c7b4a85c8de [ 604.003278] env[61906]: DEBUG nova.network.neutron [req-c579c841-c461-42c9-8142-f75485f8b868 req-f51241c5-a773-42d2-8496-d56e24a665dc service nova] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.043101] env[61906]: DEBUG nova.network.neutron [-] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.078913] env[61906]: DEBUG oslo_concurrency.lockutils [req-f8266946-2ce0-4148-a17f-ef7fdbc972ed req-bebc220c-bc39-40a8-a25f-0554e943e06a service nova] Releasing lock "refresh_cache-263da5ea-892b-4142-b487-96444abc97be" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.121688] env[61906]: DEBUG nova.network.neutron [req-c579c841-c461-42c9-8142-f75485f8b868 req-f51241c5-a773-42d2-8496-d56e24a665dc service nova] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.210358] env[61906]: DEBUG nova.compute.manager [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 604.238416] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bcd9eae-d1dd-46cd-a28e-2459daf3f3c4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.250163] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b527a4c5-2cfb-4a28-8942-b116b96544e7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.286579] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24bcd680-f6ca-4097-a017-24b0e4e56b5a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.294961] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4766c449-eeb2-4fdd-86a8-0fe85815d854 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.311155] env[61906]: DEBUG nova.compute.provider_tree [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 604.479529] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a91b120-e719-4b5f-8c70-c5037582d21c tempest-ServersWithSpecificFlavorTestJSON-2101619473 tempest-ServersWithSpecificFlavorTestJSON-2101619473-project-member] Lock "75124494-73f0-4188-817b-1c7b4a85c8de" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.633s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.545626] env[61906]: INFO nova.compute.manager [-] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Took 1.04 seconds to deallocate network for instance. [ 604.548900] env[61906]: DEBUG nova.compute.claims [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 604.549118] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.626839] env[61906]: DEBUG oslo_concurrency.lockutils [req-c579c841-c461-42c9-8142-f75485f8b868 req-f51241c5-a773-42d2-8496-d56e24a665dc service nova] Releasing lock "refresh_cache-1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.685857] env[61906]: DEBUG nova.compute.manager [req-43d7eec3-2ecd-4069-ade8-4bce6d9be228 req-fe227d05-c192-4c72-91d2-0333b62f7250 service nova] [instance: 263da5ea-892b-4142-b487-96444abc97be] Received event network-vif-deleted-e59486b6-bb9d-4623-858d-2956318d9053 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 604.817109] env[61906]: DEBUG nova.scheduler.client.report [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 604.860056] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Acquiring lock "b071a9b5-3416-4898-9ae2-ef62186c1843" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.860274] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Lock "b071a9b5-3416-4898-9ae2-ef62186c1843" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.982627] env[61906]: DEBUG nova.compute.manager [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 605.231491] env[61906]: DEBUG nova.compute.manager [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 605.267441] env[61906]: DEBUG nova.virt.hardware [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 605.267696] env[61906]: DEBUG nova.virt.hardware [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 605.267852] env[61906]: DEBUG nova.virt.hardware [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 605.268054] env[61906]: DEBUG nova.virt.hardware [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 605.268211] env[61906]: DEBUG nova.virt.hardware [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 605.269268] env[61906]: DEBUG nova.virt.hardware [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 605.269268] env[61906]: DEBUG nova.virt.hardware [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 605.269268] env[61906]: DEBUG nova.virt.hardware [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 605.269268] env[61906]: DEBUG nova.virt.hardware [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 605.269268] env[61906]: DEBUG nova.virt.hardware [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 605.269463] env[61906]: DEBUG nova.virt.hardware [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 605.270082] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f489e6b-69b8-4bf7-a238-0c3884588a22 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.280037] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68b55423-5213-4c87-ae89-b8497b9cd827 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.295143] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Instance VIF info [] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 605.300937] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Creating folder: Project (ac43352985874d48bf19b8f5df5501de). Parent ref: group-v284713. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 605.301268] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1eb5dd9b-ba7a-45ea-b3ec-5826fa1306f2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.310357] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Created folder: Project (ac43352985874d48bf19b8f5df5501de) in parent group-v284713. [ 605.310539] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Creating folder: Instances. Parent ref: group-v284721. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 605.310794] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-46b048e6-8608-4cc7-98f8-d98934ec016e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.320024] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Created folder: Instances in parent group-v284721. [ 605.320267] env[61906]: DEBUG oslo.service.loopingcall [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 605.320540] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 605.320776] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-419843ce-4cb4-4be9-a0d1-a77b46962cbf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.338048] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.154s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.338048] env[61906]: ERROR nova.compute.manager [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a196df79-d6c4-481d-b94f-7363593aabe3, please check neutron logs for more information. [ 605.338048] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Traceback (most recent call last): [ 605.338048] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 605.338048] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] self.driver.spawn(context, instance, image_meta, [ 605.338048] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 605.338048] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 605.338048] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 605.338048] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] vm_ref = self.build_virtual_machine(instance, [ 605.338423] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 605.338423] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] vif_infos = vmwarevif.get_vif_info(self._session, [ 605.338423] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 605.338423] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] for vif in network_info: [ 605.338423] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 605.338423] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] return self._sync_wrapper(fn, *args, **kwargs) [ 605.338423] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 605.338423] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] self.wait() [ 605.338423] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 605.338423] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] self[:] = self._gt.wait() [ 605.338423] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 605.338423] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] return self._exit_event.wait() [ 605.338423] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 605.338796] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] result = hub.switch() [ 605.338796] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 605.338796] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] return self.greenlet.switch() [ 605.338796] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.338796] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] result = function(*args, **kwargs) [ 605.338796] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 605.338796] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] return func(*args, **kwargs) [ 605.338796] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 605.338796] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] raise e [ 605.338796] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.338796] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] nwinfo = self.network_api.allocate_for_instance( [ 605.338796] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.338796] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] created_port_ids = self._update_ports_for_instance( [ 605.339463] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.339463] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] with excutils.save_and_reraise_exception(): [ 605.339463] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.339463] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] self.force_reraise() [ 605.339463] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.339463] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] raise self.value [ 605.339463] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.339463] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] updated_port = self._update_port( [ 605.339463] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.339463] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] _ensure_no_port_binding_failure(port) [ 605.339463] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.339463] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] raise exception.PortBindingFailed(port_id=port['id']) [ 605.340014] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] nova.exception.PortBindingFailed: Binding failed for port a196df79-d6c4-481d-b94f-7363593aabe3, please check neutron logs for more information. [ 605.340014] env[61906]: ERROR nova.compute.manager [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] [ 605.340014] env[61906]: DEBUG nova.compute.utils [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Binding failed for port a196df79-d6c4-481d-b94f-7363593aabe3, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 605.340014] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.179s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.340014] env[61906]: INFO nova.compute.claims [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 605.342983] env[61906]: DEBUG nova.compute.manager [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Build of instance da9ac1bd-97c8-4c05-9364-8bd60be50ff2 was re-scheduled: Binding failed for port a196df79-d6c4-481d-b94f-7363593aabe3, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 605.343462] env[61906]: DEBUG nova.compute.manager [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 605.343683] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Acquiring lock "refresh_cache-da9ac1bd-97c8-4c05-9364-8bd60be50ff2" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.343821] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Acquired lock "refresh_cache-da9ac1bd-97c8-4c05-9364-8bd60be50ff2" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.343973] env[61906]: DEBUG nova.network.neutron [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 605.346463] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 605.346463] env[61906]: value = "task-1333082" [ 605.346463] env[61906]: _type = "Task" [ 605.346463] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.357535] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333082, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.512832] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.812458] env[61906]: DEBUG nova.compute.manager [req-bf530415-7a57-4106-ab46-57e5a0d20f05 req-9f1212a9-b9b6-4550-ad67-8630bae9fc39 service nova] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Received event network-vif-deleted-d43156d0-1126-4b2d-8862-1fbd2a120665 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 605.867591] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333082, 'name': CreateVM_Task, 'duration_secs': 0.307706} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.868654] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 605.869374] env[61906]: DEBUG oslo_vmware.service [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-610c7194-d9e2-4a3a-90bf-d523e95094ff {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.875423] env[61906]: DEBUG nova.network.neutron [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.879664] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.879847] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.880240] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 605.883280] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d1a6399-47b8-4491-95cb-fc54d3ca51e5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.885574] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for the task: (returnval){ [ 605.885574] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5215de1e-4fa3-77f4-42fe-e8058c84a6aa" [ 605.885574] env[61906]: _type = "Task" [ 605.885574] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.893808] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5215de1e-4fa3-77f4-42fe-e8058c84a6aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.912683] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Acquiring lock "a07e49a5-e20a-49df-9326-a2e819934d15" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.912683] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Lock "a07e49a5-e20a-49df-9326-a2e819934d15" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.950011] env[61906]: DEBUG nova.network.neutron [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.399907] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.400211] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 606.400436] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.400599] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.400765] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 606.401717] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1d65fa75-e174-4f9b-a1dd-f804f6cf5384 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.423634] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 606.423857] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 606.424663] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32148651-de05-4c1f-a21a-6cf931e5cf8b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.436054] env[61906]: DEBUG oslo_concurrency.lockutils [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Acquiring lock "58b454d7-5370-499a-93dc-8613ee997b4d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.436256] env[61906]: DEBUG oslo_concurrency.lockutils [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Lock "58b454d7-5370-499a-93dc-8613ee997b4d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.439182] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6fad961-9988-457e-a33a-a2adf6cd628f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.447547] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for the task: (returnval){ [ 606.447547] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52a49c98-4a4d-6f89-59e1-22766442e331" [ 606.447547] env[61906]: _type = "Task" [ 606.447547] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.453200] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Releasing lock "refresh_cache-da9ac1bd-97c8-4c05-9364-8bd60be50ff2" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.453449] env[61906]: DEBUG nova.compute.manager [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 606.453970] env[61906]: DEBUG nova.compute.manager [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 606.454204] env[61906]: DEBUG nova.network.neutron [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 606.459824] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52a49c98-4a4d-6f89-59e1-22766442e331, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.478775] env[61906]: DEBUG nova.network.neutron [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.922993] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-830b67a8-64a9-434a-8117-696b632fddc3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.930847] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a881db1d-cf04-44a9-b510-6e3f6e5826da {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.972958] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7837b69e-9c39-4225-bf0f-7e3ef25ada6f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.982698] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Preparing fetch location {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 606.982958] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Creating directory with path [datastore1] vmware_temp/227ae18b-4317-4438-bd77-b9bb1a223177/a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 606.983315] env[61906]: DEBUG nova.network.neutron [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.984392] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-99762c9b-129c-407f-a0d9-09c44a52a1ea {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.987416] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c9b6d2-41aa-454a-9cad-652bda42b02c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.002780] env[61906]: DEBUG nova.compute.provider_tree [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 607.047171] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Created directory with path [datastore1] vmware_temp/227ae18b-4317-4438-bd77-b9bb1a223177/a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 607.047385] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Fetch image to [datastore1] vmware_temp/227ae18b-4317-4438-bd77-b9bb1a223177/a9502563-1f96-46a5-b87e-12c0e9375e6e/tmp-sparse.vmdk {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 607.048921] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Downloading image file data a9502563-1f96-46a5-b87e-12c0e9375e6e to [datastore1] vmware_temp/227ae18b-4317-4438-bd77-b9bb1a223177/a9502563-1f96-46a5-b87e-12c0e9375e6e/tmp-sparse.vmdk on the data store datastore1 {{(pid=61906) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 607.048921] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51fdff66-31a4-4d31-b3d2-5ae6d7d8c7c2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.060145] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f887a061-5654-4edd-89aa-9790dd7f1061 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.071773] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a37ea66-f950-4d50-9d8f-f467ccc63533 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.108022] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be4c57b3-5e17-45bd-a970-21095ce5097f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.113630] env[61906]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-51522f9b-46c5-4adc-aa2f-0a9f26273c27 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.137185] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Downloading image file data a9502563-1f96-46a5-b87e-12c0e9375e6e to the data store datastore1 {{(pid=61906) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 607.153899] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "3df44b70-ee72-40d7-9129-b5e94e121b0a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.154192] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "3df44b70-ee72-40d7-9129-b5e94e121b0a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.211063] env[61906]: DEBUG oslo_vmware.rw_handles [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/227ae18b-4317-4438-bd77-b9bb1a223177/a9502563-1f96-46a5-b87e-12c0e9375e6e/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61906) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 607.492671] env[61906]: INFO nova.compute.manager [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: da9ac1bd-97c8-4c05-9364-8bd60be50ff2] Took 1.04 seconds to deallocate network for instance. [ 607.508129] env[61906]: DEBUG nova.scheduler.client.report [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 607.972875] env[61906]: DEBUG oslo_vmware.rw_handles [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Completed reading data from the image iterator. {{(pid=61906) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 607.973311] env[61906]: DEBUG oslo_vmware.rw_handles [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Closing write handle for https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/227ae18b-4317-4438-bd77-b9bb1a223177/a9502563-1f96-46a5-b87e-12c0e9375e6e/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61906) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 608.013020] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.672s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.013020] env[61906]: DEBUG nova.compute.manager [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 608.014236] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.973s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.021147] env[61906]: INFO nova.compute.claims [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 608.025207] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Downloaded image file data a9502563-1f96-46a5-b87e-12c0e9375e6e to vmware_temp/227ae18b-4317-4438-bd77-b9bb1a223177/a9502563-1f96-46a5-b87e-12c0e9375e6e/tmp-sparse.vmdk on the data store datastore1 {{(pid=61906) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 608.027478] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Caching image {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 608.028027] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Copying Virtual Disk [datastore1] vmware_temp/227ae18b-4317-4438-bd77-b9bb1a223177/a9502563-1f96-46a5-b87e-12c0e9375e6e/tmp-sparse.vmdk to [datastore1] vmware_temp/227ae18b-4317-4438-bd77-b9bb1a223177/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 608.030017] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e2e12fb0-eb96-4506-9463-274815618dd3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.038315] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for the task: (returnval){ [ 608.038315] env[61906]: value = "task-1333083" [ 608.038315] env[61906]: _type = "Task" [ 608.038315] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.049299] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333083, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.534334] env[61906]: DEBUG nova.compute.utils [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 608.538303] env[61906]: DEBUG nova.compute.manager [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 608.538441] env[61906]: DEBUG nova.network.neutron [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 608.542072] env[61906]: INFO nova.scheduler.client.report [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Deleted allocations for instance da9ac1bd-97c8-4c05-9364-8bd60be50ff2 [ 608.562162] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333083, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.624876] env[61906]: DEBUG nova.policy [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '670a8b85f9774262841841b30b9005cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8cb9315c34724161a787b949e5976800', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 609.044308] env[61906]: DEBUG nova.compute.manager [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 609.058798] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8dabc030-9442-4189-a8d4-7d5ab4800c3e tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Lock "da9ac1bd-97c8-4c05-9364-8bd60be50ff2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.427s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.067938] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333083, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.672221} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.068261] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Copied Virtual Disk [datastore1] vmware_temp/227ae18b-4317-4438-bd77-b9bb1a223177/a9502563-1f96-46a5-b87e-12c0e9375e6e/tmp-sparse.vmdk to [datastore1] vmware_temp/227ae18b-4317-4438-bd77-b9bb1a223177/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 609.068447] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Deleting the datastore file [datastore1] vmware_temp/227ae18b-4317-4438-bd77-b9bb1a223177/a9502563-1f96-46a5-b87e-12c0e9375e6e/tmp-sparse.vmdk {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 609.068711] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-31b92a8f-9c8c-4a6b-8f46-1de7d82adc8e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.075537] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for the task: (returnval){ [ 609.075537] env[61906]: value = "task-1333084" [ 609.075537] env[61906]: _type = "Task" [ 609.075537] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.085501] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333084, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.110050] env[61906]: DEBUG nova.network.neutron [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Successfully created port: fa2b5a1c-551c-44cc-bbf8-8f23c6f3e8ab {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 609.560366] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 609.594103] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333084, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025752} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.599356] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 609.599742] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Moving file from [datastore1] vmware_temp/227ae18b-4317-4438-bd77-b9bb1a223177/a9502563-1f96-46a5-b87e-12c0e9375e6e to [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e. {{(pid=61906) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 609.605754] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-1af35166-4171-46d4-a408-f06e67243cbe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.614957] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for the task: (returnval){ [ 609.614957] env[61906]: value = "task-1333085" [ 609.614957] env[61906]: _type = "Task" [ 609.614957] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.632410] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333085, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.641514] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2855074-0308-4938-969e-0fa6adad978e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.652590] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19be73f2-3e85-4f51-8e2a-3d72388da5b7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.692778] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e69fc9bb-d0a5-4208-86a0-bf43c1d2b9a6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.699427] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-760f1e5a-04d7-42dd-8fbc-28f6cf381bc9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.712266] env[61906]: DEBUG nova.compute.provider_tree [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.056117] env[61906]: DEBUG nova.compute.manager [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 610.085150] env[61906]: DEBUG nova.virt.hardware [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 610.085470] env[61906]: DEBUG nova.virt.hardware [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 610.085632] env[61906]: DEBUG nova.virt.hardware [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 610.085993] env[61906]: DEBUG nova.virt.hardware [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 610.086169] env[61906]: DEBUG nova.virt.hardware [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 610.086318] env[61906]: DEBUG nova.virt.hardware [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 610.086645] env[61906]: DEBUG nova.virt.hardware [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 610.086818] env[61906]: DEBUG nova.virt.hardware [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 610.087051] env[61906]: DEBUG nova.virt.hardware [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 610.087222] env[61906]: DEBUG nova.virt.hardware [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 610.087389] env[61906]: DEBUG nova.virt.hardware [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 610.088246] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd984f03-241d-4845-849b-04d1db7fb478 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.091564] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.099273] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7de61e4a-0117-4a75-8651-18ac100b15cd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.124636] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333085, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.05398} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.124636] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] File moved {{(pid=61906) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 610.124636] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Cleaning up location [datastore1] vmware_temp/227ae18b-4317-4438-bd77-b9bb1a223177 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 610.124636] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Deleting the datastore file [datastore1] vmware_temp/227ae18b-4317-4438-bd77-b9bb1a223177 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 610.124636] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-317e578c-cb39-43ee-be43-f42699586a59 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.130720] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for the task: (returnval){ [ 610.130720] env[61906]: value = "task-1333086" [ 610.130720] env[61906]: _type = "Task" [ 610.130720] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.140019] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333086, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.215958] env[61906]: DEBUG nova.scheduler.client.report [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 610.458271] env[61906]: DEBUG nova.compute.manager [req-62dec8df-c451-4936-a19b-93c9e47b87de req-a792b6ce-3500-4b8c-af4f-90677b51654b service nova] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Received event network-changed-fa2b5a1c-551c-44cc-bbf8-8f23c6f3e8ab {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 610.458479] env[61906]: DEBUG nova.compute.manager [req-62dec8df-c451-4936-a19b-93c9e47b87de req-a792b6ce-3500-4b8c-af4f-90677b51654b service nova] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Refreshing instance network info cache due to event network-changed-fa2b5a1c-551c-44cc-bbf8-8f23c6f3e8ab. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 610.459132] env[61906]: DEBUG oslo_concurrency.lockutils [req-62dec8df-c451-4936-a19b-93c9e47b87de req-a792b6ce-3500-4b8c-af4f-90677b51654b service nova] Acquiring lock "refresh_cache-7ed5b09c-5ba1-479c-96b4-75924df84e75" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.459132] env[61906]: DEBUG oslo_concurrency.lockutils [req-62dec8df-c451-4936-a19b-93c9e47b87de req-a792b6ce-3500-4b8c-af4f-90677b51654b service nova] Acquired lock "refresh_cache-7ed5b09c-5ba1-479c-96b4-75924df84e75" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.459132] env[61906]: DEBUG nova.network.neutron [req-62dec8df-c451-4936-a19b-93c9e47b87de req-a792b6ce-3500-4b8c-af4f-90677b51654b service nova] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Refreshing network info cache for port fa2b5a1c-551c-44cc-bbf8-8f23c6f3e8ab {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 610.641432] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333086, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.022064} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.642601] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 610.643022] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7feab5fc-3a1f-4d6d-8d67-ea5545666f1c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.648587] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for the task: (returnval){ [ 610.648587] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5222fb45-796f-5b5f-bc5f-86296be91a9e" [ 610.648587] env[61906]: _type = "Task" [ 610.648587] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.660222] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5222fb45-796f-5b5f-bc5f-86296be91a9e, 'name': SearchDatastore_Task, 'duration_secs': 0.009189} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.660467] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.660756] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 318a3224-a156-4354-a73b-460168c0eb9c/318a3224-a156-4354-a73b-460168c0eb9c.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 610.661019] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7c1b4cc8-92e5-4b93-820e-815d55ecfb6f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.667358] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for the task: (returnval){ [ 610.667358] env[61906]: value = "task-1333087" [ 610.667358] env[61906]: _type = "Task" [ 610.667358] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.675595] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333087, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.676570] env[61906]: ERROR nova.compute.manager [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fa2b5a1c-551c-44cc-bbf8-8f23c6f3e8ab, please check neutron logs for more information. [ 610.676570] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 610.676570] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.676570] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 610.676570] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.676570] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 610.676570] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.676570] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 610.676570] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.676570] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 610.676570] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.676570] env[61906]: ERROR nova.compute.manager raise self.value [ 610.676570] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.676570] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 610.676570] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.676570] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 610.677168] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.677168] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 610.677168] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fa2b5a1c-551c-44cc-bbf8-8f23c6f3e8ab, please check neutron logs for more information. [ 610.677168] env[61906]: ERROR nova.compute.manager [ 610.677168] env[61906]: Traceback (most recent call last): [ 610.677168] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 610.677168] env[61906]: listener.cb(fileno) [ 610.677168] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.677168] env[61906]: result = function(*args, **kwargs) [ 610.677168] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 610.677168] env[61906]: return func(*args, **kwargs) [ 610.677168] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.677168] env[61906]: raise e [ 610.677168] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.677168] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 610.677168] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.677168] env[61906]: created_port_ids = self._update_ports_for_instance( [ 610.677168] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.677168] env[61906]: with excutils.save_and_reraise_exception(): [ 610.677168] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.677168] env[61906]: self.force_reraise() [ 610.677168] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.677168] env[61906]: raise self.value [ 610.677168] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.677168] env[61906]: updated_port = self._update_port( [ 610.677168] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.677168] env[61906]: _ensure_no_port_binding_failure(port) [ 610.677168] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.677168] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 610.678027] env[61906]: nova.exception.PortBindingFailed: Binding failed for port fa2b5a1c-551c-44cc-bbf8-8f23c6f3e8ab, please check neutron logs for more information. [ 610.678027] env[61906]: Removing descriptor: 16 [ 610.678027] env[61906]: ERROR nova.compute.manager [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fa2b5a1c-551c-44cc-bbf8-8f23c6f3e8ab, please check neutron logs for more information. [ 610.678027] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Traceback (most recent call last): [ 610.678027] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 610.678027] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] yield resources [ 610.678027] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 610.678027] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] self.driver.spawn(context, instance, image_meta, [ 610.678027] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 610.678027] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] self._vmops.spawn(context, instance, image_meta, injected_files, [ 610.678027] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 610.678027] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] vm_ref = self.build_virtual_machine(instance, [ 610.678455] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 610.678455] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] vif_infos = vmwarevif.get_vif_info(self._session, [ 610.678455] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 610.678455] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] for vif in network_info: [ 610.678455] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 610.678455] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] return self._sync_wrapper(fn, *args, **kwargs) [ 610.678455] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 610.678455] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] self.wait() [ 610.678455] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 610.678455] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] self[:] = self._gt.wait() [ 610.678455] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 610.678455] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] return self._exit_event.wait() [ 610.678455] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 610.678983] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] result = hub.switch() [ 610.678983] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 610.678983] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] return self.greenlet.switch() [ 610.678983] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.678983] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] result = function(*args, **kwargs) [ 610.678983] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 610.678983] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] return func(*args, **kwargs) [ 610.678983] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.678983] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] raise e [ 610.678983] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.678983] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] nwinfo = self.network_api.allocate_for_instance( [ 610.678983] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.678983] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] created_port_ids = self._update_ports_for_instance( [ 610.679432] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.679432] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] with excutils.save_and_reraise_exception(): [ 610.679432] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.679432] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] self.force_reraise() [ 610.679432] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.679432] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] raise self.value [ 610.679432] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.679432] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] updated_port = self._update_port( [ 610.679432] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.679432] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] _ensure_no_port_binding_failure(port) [ 610.679432] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.679432] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] raise exception.PortBindingFailed(port_id=port['id']) [ 610.680062] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] nova.exception.PortBindingFailed: Binding failed for port fa2b5a1c-551c-44cc-bbf8-8f23c6f3e8ab, please check neutron logs for more information. [ 610.680062] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] [ 610.680062] env[61906]: INFO nova.compute.manager [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Terminating instance [ 610.680062] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Acquiring lock "refresh_cache-7ed5b09c-5ba1-479c-96b4-75924df84e75" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.723552] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.709s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.724165] env[61906]: DEBUG nova.compute.manager [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 610.727326] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.662s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.728884] env[61906]: INFO nova.compute.claims [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 610.986134] env[61906]: DEBUG nova.network.neutron [req-62dec8df-c451-4936-a19b-93c9e47b87de req-a792b6ce-3500-4b8c-af4f-90677b51654b service nova] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.101294] env[61906]: DEBUG nova.network.neutron [req-62dec8df-c451-4936-a19b-93c9e47b87de req-a792b6ce-3500-4b8c-af4f-90677b51654b service nova] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.177994] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333087, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.469405} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.178361] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 318a3224-a156-4354-a73b-460168c0eb9c/318a3224-a156-4354-a73b-460168c0eb9c.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 611.178577] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 611.178962] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7dc56629-122a-4363-bfe0-fd5da4d5e3c6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.184879] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for the task: (returnval){ [ 611.184879] env[61906]: value = "task-1333088" [ 611.184879] env[61906]: _type = "Task" [ 611.184879] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.194954] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333088, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.238018] env[61906]: DEBUG nova.compute.utils [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 611.238273] env[61906]: DEBUG nova.compute.manager [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 611.238723] env[61906]: DEBUG nova.network.neutron [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 611.309428] env[61906]: DEBUG nova.policy [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e89d928ccea04803b8e0ecf71369b44c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6ca2ae9c0a3c45e6a284bdffbf5198f6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 611.607354] env[61906]: DEBUG oslo_concurrency.lockutils [req-62dec8df-c451-4936-a19b-93c9e47b87de req-a792b6ce-3500-4b8c-af4f-90677b51654b service nova] Releasing lock "refresh_cache-7ed5b09c-5ba1-479c-96b4-75924df84e75" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.607691] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Acquired lock "refresh_cache-7ed5b09c-5ba1-479c-96b4-75924df84e75" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.609041] env[61906]: DEBUG nova.network.neutron [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 611.697980] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333088, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059469} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.699176] env[61906]: DEBUG nova.network.neutron [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Successfully created port: c43c348f-e3b1-48e5-911e-643fcc5dcef6 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 611.700692] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 611.701537] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeaeffbd-830b-4f58-a1f5-8e3c9fa4002a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.723578] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Reconfiguring VM instance instance-0000000d to attach disk [datastore1] 318a3224-a156-4354-a73b-460168c0eb9c/318a3224-a156-4354-a73b-460168c0eb9c.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 611.723861] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d34370f9-4504-4a9e-b9d2-5e9fa2921152 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.742861] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for the task: (returnval){ [ 611.742861] env[61906]: value = "task-1333089" [ 611.742861] env[61906]: _type = "Task" [ 611.742861] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.746099] env[61906]: DEBUG nova.compute.manager [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 611.756568] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333089, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.129491] env[61906]: DEBUG nova.network.neutron [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.260996] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333089, 'name': ReconfigVM_Task, 'duration_secs': 0.280935} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.262938] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Reconfigured VM instance instance-0000000d to attach disk [datastore1] 318a3224-a156-4354-a73b-460168c0eb9c/318a3224-a156-4354-a73b-460168c0eb9c.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 612.263706] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-78212684-8a9a-4ca9-a1e0-230d9d4f367f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.271280] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for the task: (returnval){ [ 612.271280] env[61906]: value = "task-1333090" [ 612.271280] env[61906]: _type = "Task" [ 612.271280] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.279446] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333090, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.307207] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34907d2a-8614-4760-a1e9-cbc4964ac240 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.313912] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d8d8531-473f-4d4a-8834-d9ab1f9e6528 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.348066] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d20522-1969-4935-a4e6-c184fad11dff {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.356168] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5715e4e-787c-4c71-9c8b-f1b100345412 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.375020] env[61906]: DEBUG nova.compute.provider_tree [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 612.407350] env[61906]: DEBUG nova.network.neutron [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.599825] env[61906]: DEBUG nova.compute.manager [req-0a285717-60dd-4086-ade2-6d757a9dca64 req-8145ebd6-523e-4ef0-b79b-aaa6a68d9a94 service nova] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Received event network-vif-deleted-fa2b5a1c-551c-44cc-bbf8-8f23c6f3e8ab {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 612.766736] env[61906]: DEBUG nova.compute.manager [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 612.782159] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333090, 'name': Rename_Task, 'duration_secs': 0.134097} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.782437] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 612.782672] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3f16104a-745f-4d52-8dd7-689b4cdb668a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.788733] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for the task: (returnval){ [ 612.788733] env[61906]: value = "task-1333091" [ 612.788733] env[61906]: _type = "Task" [ 612.788733] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.804438] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333091, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.808099] env[61906]: DEBUG nova.virt.hardware [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 612.808099] env[61906]: DEBUG nova.virt.hardware [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 612.808099] env[61906]: DEBUG nova.virt.hardware [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 612.808376] env[61906]: DEBUG nova.virt.hardware [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 612.808376] env[61906]: DEBUG nova.virt.hardware [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 612.808376] env[61906]: DEBUG nova.virt.hardware [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 612.808376] env[61906]: DEBUG nova.virt.hardware [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 612.808376] env[61906]: DEBUG nova.virt.hardware [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 612.808800] env[61906]: DEBUG nova.virt.hardware [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 612.808800] env[61906]: DEBUG nova.virt.hardware [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 612.808800] env[61906]: DEBUG nova.virt.hardware [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 612.810741] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5547e7aa-eeb3-493e-b652-6fd1c78e1685 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.818938] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caf4591a-e2aa-4631-8b49-55f54bd1bbfd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.878524] env[61906]: DEBUG nova.scheduler.client.report [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 612.909416] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Releasing lock "refresh_cache-7ed5b09c-5ba1-479c-96b4-75924df84e75" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.909930] env[61906]: DEBUG nova.compute.manager [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 612.910160] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 612.910460] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5be5890a-d53a-4696-8d91-f6bd80a0716e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.919718] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7805d8de-b20c-4773-b4d3-1da604d85cea {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.944966] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7ed5b09c-5ba1-479c-96b4-75924df84e75 could not be found. [ 612.944966] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 612.944966] env[61906]: INFO nova.compute.manager [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Took 0.03 seconds to destroy the instance on the hypervisor. [ 612.945182] env[61906]: DEBUG oslo.service.loopingcall [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 612.945283] env[61906]: DEBUG nova.compute.manager [-] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 612.945322] env[61906]: DEBUG nova.network.neutron [-] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 612.967659] env[61906]: DEBUG nova.network.neutron [-] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.013220] env[61906]: ERROR nova.compute.manager [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c43c348f-e3b1-48e5-911e-643fcc5dcef6, please check neutron logs for more information. [ 613.013220] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 613.013220] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.013220] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 613.013220] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 613.013220] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 613.013220] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 613.013220] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 613.013220] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.013220] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 613.013220] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.013220] env[61906]: ERROR nova.compute.manager raise self.value [ 613.013220] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 613.013220] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 613.013220] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.013220] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 613.015072] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.015072] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 613.015072] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c43c348f-e3b1-48e5-911e-643fcc5dcef6, please check neutron logs for more information. [ 613.015072] env[61906]: ERROR nova.compute.manager [ 613.015072] env[61906]: Traceback (most recent call last): [ 613.015072] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 613.015072] env[61906]: listener.cb(fileno) [ 613.015072] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.015072] env[61906]: result = function(*args, **kwargs) [ 613.015072] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 613.015072] env[61906]: return func(*args, **kwargs) [ 613.015072] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.015072] env[61906]: raise e [ 613.015072] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.015072] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 613.015072] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 613.015072] env[61906]: created_port_ids = self._update_ports_for_instance( [ 613.015072] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 613.015072] env[61906]: with excutils.save_and_reraise_exception(): [ 613.015072] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.015072] env[61906]: self.force_reraise() [ 613.015072] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.015072] env[61906]: raise self.value [ 613.015072] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 613.015072] env[61906]: updated_port = self._update_port( [ 613.015072] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.015072] env[61906]: _ensure_no_port_binding_failure(port) [ 613.015072] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.015072] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 613.016912] env[61906]: nova.exception.PortBindingFailed: Binding failed for port c43c348f-e3b1-48e5-911e-643fcc5dcef6, please check neutron logs for more information. [ 613.016912] env[61906]: Removing descriptor: 16 [ 613.016912] env[61906]: ERROR nova.compute.manager [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c43c348f-e3b1-48e5-911e-643fcc5dcef6, please check neutron logs for more information. [ 613.016912] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Traceback (most recent call last): [ 613.016912] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 613.016912] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] yield resources [ 613.016912] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 613.016912] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] self.driver.spawn(context, instance, image_meta, [ 613.016912] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 613.016912] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] self._vmops.spawn(context, instance, image_meta, injected_files, [ 613.016912] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 613.016912] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] vm_ref = self.build_virtual_machine(instance, [ 613.017797] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 613.017797] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] vif_infos = vmwarevif.get_vif_info(self._session, [ 613.017797] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 613.017797] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] for vif in network_info: [ 613.017797] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 613.017797] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] return self._sync_wrapper(fn, *args, **kwargs) [ 613.017797] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 613.017797] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] self.wait() [ 613.017797] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 613.017797] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] self[:] = self._gt.wait() [ 613.017797] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 613.017797] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] return self._exit_event.wait() [ 613.017797] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 613.018619] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] result = hub.switch() [ 613.018619] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 613.018619] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] return self.greenlet.switch() [ 613.018619] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.018619] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] result = function(*args, **kwargs) [ 613.018619] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 613.018619] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] return func(*args, **kwargs) [ 613.018619] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.018619] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] raise e [ 613.018619] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.018619] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] nwinfo = self.network_api.allocate_for_instance( [ 613.018619] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 613.018619] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] created_port_ids = self._update_ports_for_instance( [ 613.019255] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 613.019255] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] with excutils.save_and_reraise_exception(): [ 613.019255] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.019255] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] self.force_reraise() [ 613.019255] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.019255] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] raise self.value [ 613.019255] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 613.019255] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] updated_port = self._update_port( [ 613.019255] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.019255] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] _ensure_no_port_binding_failure(port) [ 613.019255] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.019255] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] raise exception.PortBindingFailed(port_id=port['id']) [ 613.019956] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] nova.exception.PortBindingFailed: Binding failed for port c43c348f-e3b1-48e5-911e-643fcc5dcef6, please check neutron logs for more information. [ 613.019956] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] [ 613.019956] env[61906]: INFO nova.compute.manager [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Terminating instance [ 613.020536] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Acquiring lock "refresh_cache-a9249055-f6bf-4693-b0f7-5d8b63510a35" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.020702] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Acquired lock "refresh_cache-a9249055-f6bf-4693-b0f7-5d8b63510a35" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.020885] env[61906]: DEBUG nova.network.neutron [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 613.299161] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333091, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.383224] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.656s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.383745] env[61906]: DEBUG nova.compute.manager [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 613.386493] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.115s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.388689] env[61906]: INFO nova.compute.claims [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 613.470255] env[61906]: DEBUG nova.network.neutron [-] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.548375] env[61906]: DEBUG nova.network.neutron [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.653678] env[61906]: DEBUG nova.network.neutron [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.800700] env[61906]: DEBUG oslo_vmware.api [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333091, 'name': PowerOnVM_Task, 'duration_secs': 0.777744} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.800992] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 613.801223] env[61906]: INFO nova.compute.manager [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Took 8.57 seconds to spawn the instance on the hypervisor. [ 613.801367] env[61906]: DEBUG nova.compute.manager [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 613.802425] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0ddddab-a638-4127-8371-14043c33f928 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.900044] env[61906]: DEBUG nova.compute.utils [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 613.904103] env[61906]: DEBUG nova.compute.manager [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 613.904287] env[61906]: DEBUG nova.network.neutron [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 613.963139] env[61906]: DEBUG nova.policy [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bf22a6ff9c8541b59ba9a72529c48027', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5972cff43b0b4d5ab21ed30f126d94ef', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 613.975326] env[61906]: INFO nova.compute.manager [-] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Took 1.03 seconds to deallocate network for instance. [ 613.977500] env[61906]: DEBUG nova.compute.claims [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 613.977737] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.162688] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Releasing lock "refresh_cache-a9249055-f6bf-4693-b0f7-5d8b63510a35" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.163145] env[61906]: DEBUG nova.compute.manager [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 614.163360] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 614.163699] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9b137932-74a6-4af3-8913-e4701e6dd42d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.178166] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50bdf0f3-7e98-4db3-91ff-8c95ab28712a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.204812] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a9249055-f6bf-4693-b0f7-5d8b63510a35 could not be found. [ 614.205638] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 614.205638] env[61906]: INFO nova.compute.manager [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Took 0.04 seconds to destroy the instance on the hypervisor. [ 614.205638] env[61906]: DEBUG oslo.service.loopingcall [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 614.206124] env[61906]: DEBUG nova.compute.manager [-] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 614.206247] env[61906]: DEBUG nova.network.neutron [-] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 614.233780] env[61906]: DEBUG nova.network.neutron [-] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 614.324336] env[61906]: INFO nova.compute.manager [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Took 30.08 seconds to build instance. [ 614.340445] env[61906]: DEBUG nova.compute.manager [None req-24137690-26da-48d7-91c4-7d5f3ce9b852 tempest-ServerDiagnosticsV248Test-447830524 tempest-ServerDiagnosticsV248Test-447830524-project-admin] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 614.341634] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8368d43-315d-436f-9cbb-99381efaf6e4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.346117] env[61906]: DEBUG nova.network.neutron [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Successfully created port: 0206d2c8-6ff4-4df4-a6e2-b6a3d7e4c46a {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 614.354918] env[61906]: INFO nova.compute.manager [None req-24137690-26da-48d7-91c4-7d5f3ce9b852 tempest-ServerDiagnosticsV248Test-447830524 tempest-ServerDiagnosticsV248Test-447830524-project-admin] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Retrieving diagnostics [ 614.355787] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43bf6a95-fa7a-4d02-afed-b7e94fa68eb3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.405765] env[61906]: DEBUG nova.compute.manager [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 614.722719] env[61906]: INFO nova.compute.manager [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Rebuilding instance [ 614.737181] env[61906]: DEBUG nova.network.neutron [-] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.776149] env[61906]: DEBUG nova.compute.manager [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 614.778378] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-413d592a-2720-4609-b22b-14c85286362a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.825106] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fbbf0609-0bbf-45bd-8e42-370251c3cfcc tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Lock "318a3224-a156-4354-a73b-460168c0eb9c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.817s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.058178] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-397bffee-8317-4d11-a80e-c14fbefffa42 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.065432] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1bac86a-04ca-4272-aa2c-900ad220987e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.096586] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc25f8d-9dd8-4816-b3c1-9c51eaaa0ec4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.104024] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d54c68-dd4e-4ffa-9b19-d16bc4d0193b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.117534] env[61906]: DEBUG nova.compute.provider_tree [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.242673] env[61906]: INFO nova.compute.manager [-] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Took 1.04 seconds to deallocate network for instance. [ 615.245601] env[61906]: DEBUG nova.compute.claims [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 615.245798] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.292667] env[61906]: DEBUG nova.compute.manager [req-96fa4fc1-d46d-4cd0-878f-0593e4f9d94a req-d30784e1-ff54-4a81-8cba-b4feee1e67c5 service nova] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Received event network-changed-c43c348f-e3b1-48e5-911e-643fcc5dcef6 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 615.292943] env[61906]: DEBUG nova.compute.manager [req-96fa4fc1-d46d-4cd0-878f-0593e4f9d94a req-d30784e1-ff54-4a81-8cba-b4feee1e67c5 service nova] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Refreshing instance network info cache due to event network-changed-c43c348f-e3b1-48e5-911e-643fcc5dcef6. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 615.293246] env[61906]: DEBUG oslo_concurrency.lockutils [req-96fa4fc1-d46d-4cd0-878f-0593e4f9d94a req-d30784e1-ff54-4a81-8cba-b4feee1e67c5 service nova] Acquiring lock "refresh_cache-a9249055-f6bf-4693-b0f7-5d8b63510a35" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.293464] env[61906]: DEBUG oslo_concurrency.lockutils [req-96fa4fc1-d46d-4cd0-878f-0593e4f9d94a req-d30784e1-ff54-4a81-8cba-b4feee1e67c5 service nova] Acquired lock "refresh_cache-a9249055-f6bf-4693-b0f7-5d8b63510a35" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.293686] env[61906]: DEBUG nova.network.neutron [req-96fa4fc1-d46d-4cd0-878f-0593e4f9d94a req-d30784e1-ff54-4a81-8cba-b4feee1e67c5 service nova] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Refreshing network info cache for port c43c348f-e3b1-48e5-911e-643fcc5dcef6 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 615.296908] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 615.298058] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-23356b2c-8896-40c4-b58f-7c5266f624d6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.308858] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for the task: (returnval){ [ 615.308858] env[61906]: value = "task-1333092" [ 615.308858] env[61906]: _type = "Task" [ 615.308858] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.319971] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333092, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.332067] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 615.422507] env[61906]: DEBUG nova.compute.manager [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 615.453726] env[61906]: DEBUG nova.virt.hardware [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 615.453967] env[61906]: DEBUG nova.virt.hardware [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 615.454137] env[61906]: DEBUG nova.virt.hardware [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 615.454324] env[61906]: DEBUG nova.virt.hardware [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 615.454469] env[61906]: DEBUG nova.virt.hardware [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 615.454615] env[61906]: DEBUG nova.virt.hardware [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 615.454822] env[61906]: DEBUG nova.virt.hardware [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 615.454981] env[61906]: DEBUG nova.virt.hardware [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 615.455159] env[61906]: DEBUG nova.virt.hardware [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 615.455321] env[61906]: DEBUG nova.virt.hardware [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 615.455480] env[61906]: DEBUG nova.virt.hardware [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 615.456362] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6bdfade-c995-4ae5-a6d6-5b1bb2ffd8bf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.464380] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6da3e79b-0ceb-45e7-a1d0-737f2d83df49 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.621327] env[61906]: DEBUG nova.scheduler.client.report [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 615.647962] env[61906]: ERROR nova.compute.manager [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0206d2c8-6ff4-4df4-a6e2-b6a3d7e4c46a, please check neutron logs for more information. [ 615.647962] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 615.647962] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.647962] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 615.647962] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 615.647962] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 615.647962] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 615.647962] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 615.647962] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.647962] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 615.647962] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.647962] env[61906]: ERROR nova.compute.manager raise self.value [ 615.647962] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 615.647962] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 615.647962] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.647962] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 615.648580] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.648580] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 615.648580] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0206d2c8-6ff4-4df4-a6e2-b6a3d7e4c46a, please check neutron logs for more information. [ 615.648580] env[61906]: ERROR nova.compute.manager [ 615.648580] env[61906]: Traceback (most recent call last): [ 615.648580] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 615.648580] env[61906]: listener.cb(fileno) [ 615.648580] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.648580] env[61906]: result = function(*args, **kwargs) [ 615.648580] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 615.648580] env[61906]: return func(*args, **kwargs) [ 615.648580] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 615.648580] env[61906]: raise e [ 615.648580] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.648580] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 615.648580] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 615.648580] env[61906]: created_port_ids = self._update_ports_for_instance( [ 615.648580] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 615.648580] env[61906]: with excutils.save_and_reraise_exception(): [ 615.648580] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.648580] env[61906]: self.force_reraise() [ 615.648580] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.648580] env[61906]: raise self.value [ 615.648580] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 615.648580] env[61906]: updated_port = self._update_port( [ 615.648580] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.648580] env[61906]: _ensure_no_port_binding_failure(port) [ 615.648580] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.648580] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 615.649589] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 0206d2c8-6ff4-4df4-a6e2-b6a3d7e4c46a, please check neutron logs for more information. [ 615.649589] env[61906]: Removing descriptor: 16 [ 615.649589] env[61906]: ERROR nova.compute.manager [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0206d2c8-6ff4-4df4-a6e2-b6a3d7e4c46a, please check neutron logs for more information. [ 615.649589] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Traceback (most recent call last): [ 615.649589] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 615.649589] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] yield resources [ 615.649589] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 615.649589] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] self.driver.spawn(context, instance, image_meta, [ 615.649589] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 615.649589] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 615.649589] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 615.649589] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] vm_ref = self.build_virtual_machine(instance, [ 615.650081] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 615.650081] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] vif_infos = vmwarevif.get_vif_info(self._session, [ 615.650081] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 615.650081] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] for vif in network_info: [ 615.650081] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 615.650081] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] return self._sync_wrapper(fn, *args, **kwargs) [ 615.650081] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 615.650081] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] self.wait() [ 615.650081] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 615.650081] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] self[:] = self._gt.wait() [ 615.650081] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 615.650081] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] return self._exit_event.wait() [ 615.650081] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 615.650530] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] result = hub.switch() [ 615.650530] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 615.650530] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] return self.greenlet.switch() [ 615.650530] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.650530] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] result = function(*args, **kwargs) [ 615.650530] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 615.650530] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] return func(*args, **kwargs) [ 615.650530] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 615.650530] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] raise e [ 615.650530] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.650530] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] nwinfo = self.network_api.allocate_for_instance( [ 615.650530] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 615.650530] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] created_port_ids = self._update_ports_for_instance( [ 615.650974] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 615.650974] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] with excutils.save_and_reraise_exception(): [ 615.650974] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.650974] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] self.force_reraise() [ 615.650974] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.650974] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] raise self.value [ 615.650974] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 615.650974] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] updated_port = self._update_port( [ 615.650974] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.650974] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] _ensure_no_port_binding_failure(port) [ 615.650974] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.650974] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] raise exception.PortBindingFailed(port_id=port['id']) [ 615.651431] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] nova.exception.PortBindingFailed: Binding failed for port 0206d2c8-6ff4-4df4-a6e2-b6a3d7e4c46a, please check neutron logs for more information. [ 615.651431] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] [ 615.651431] env[61906]: INFO nova.compute.manager [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Terminating instance [ 615.653112] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Acquiring lock "refresh_cache-bff5e099-f6ad-41bd-a55a-6987310c37cf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.653239] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Acquired lock "refresh_cache-bff5e099-f6ad-41bd-a55a-6987310c37cf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.653407] env[61906]: DEBUG nova.network.neutron [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 615.768193] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Acquiring lock "a048328a-d4a8-420e-9cc8-047254d7d1bf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.768193] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Lock "a048328a-d4a8-420e-9cc8-047254d7d1bf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.810796] env[61906]: DEBUG nova.network.neutron [req-96fa4fc1-d46d-4cd0-878f-0593e4f9d94a req-d30784e1-ff54-4a81-8cba-b4feee1e67c5 service nova] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.822165] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333092, 'name': PowerOffVM_Task, 'duration_secs': 0.195869} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.822497] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 615.822717] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 615.823556] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cff0565-957f-44b1-808c-5212ce78644f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.831233] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 615.831846] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d99b5dbe-1056-4438-94d5-a2fa9d27d44c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.852972] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.865433] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 615.865652] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 615.865828] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Deleting the datastore file [datastore1] 318a3224-a156-4354-a73b-460168c0eb9c {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 615.866101] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-72e1c1f0-cfc1-4131-b776-a772053f8ce6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.873559] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for the task: (returnval){ [ 615.873559] env[61906]: value = "task-1333094" [ 615.873559] env[61906]: _type = "Task" [ 615.873559] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.887104] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333094, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.911722] env[61906]: DEBUG nova.network.neutron [req-96fa4fc1-d46d-4cd0-878f-0593e4f9d94a req-d30784e1-ff54-4a81-8cba-b4feee1e67c5 service nova] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.094240] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Acquiring lock "eb4cac2e-85d8-46b6-a4d4-ff12ee70e435" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.098019] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Lock "eb4cac2e-85d8-46b6-a4d4-ff12ee70e435" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.098019] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Acquiring lock "eb4cac2e-85d8-46b6-a4d4-ff12ee70e435-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.098019] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Lock "eb4cac2e-85d8-46b6-a4d4-ff12ee70e435-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.098019] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Lock "eb4cac2e-85d8-46b6-a4d4-ff12ee70e435-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.099043] env[61906]: INFO nova.compute.manager [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Terminating instance [ 616.100689] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Acquiring lock "refresh_cache-eb4cac2e-85d8-46b6-a4d4-ff12ee70e435" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.100865] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Acquired lock "refresh_cache-eb4cac2e-85d8-46b6-a4d4-ff12ee70e435" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.101042] env[61906]: DEBUG nova.network.neutron [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 616.127045] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.740s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.127045] env[61906]: DEBUG nova.compute.manager [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 616.129798] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.975s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.131308] env[61906]: INFO nova.compute.claims [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 616.172385] env[61906]: DEBUG nova.network.neutron [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 616.218126] env[61906]: DEBUG nova.network.neutron [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.387688] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333094, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.099065} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.387961] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 616.388185] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 616.388372] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 616.417136] env[61906]: DEBUG oslo_concurrency.lockutils [req-96fa4fc1-d46d-4cd0-878f-0593e4f9d94a req-d30784e1-ff54-4a81-8cba-b4feee1e67c5 service nova] Releasing lock "refresh_cache-a9249055-f6bf-4693-b0f7-5d8b63510a35" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.417136] env[61906]: DEBUG nova.compute.manager [req-96fa4fc1-d46d-4cd0-878f-0593e4f9d94a req-d30784e1-ff54-4a81-8cba-b4feee1e67c5 service nova] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Received event network-vif-deleted-c43c348f-e3b1-48e5-911e-643fcc5dcef6 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 616.617865] env[61906]: DEBUG nova.network.neutron [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 616.635777] env[61906]: DEBUG nova.compute.utils [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 616.640274] env[61906]: DEBUG nova.compute.manager [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 616.640274] env[61906]: DEBUG nova.network.neutron [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 616.676949] env[61906]: DEBUG nova.network.neutron [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.679238] env[61906]: DEBUG nova.policy [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ec5c3ff023c14611822544609687cd04', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a20dac7162a74af5a071b082bd5e5f16', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 616.720327] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Releasing lock "refresh_cache-bff5e099-f6ad-41bd-a55a-6987310c37cf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.720705] env[61906]: DEBUG nova.compute.manager [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 616.720947] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 616.721531] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-567f61ee-b479-4cee-a20e-5bb355239839 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.730364] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-084e8f82-7861-41f7-8705-c5aa3b38c892 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.754106] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bff5e099-f6ad-41bd-a55a-6987310c37cf could not be found. [ 616.754106] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 616.754106] env[61906]: INFO nova.compute.manager [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Took 0.03 seconds to destroy the instance on the hypervisor. [ 616.754106] env[61906]: DEBUG oslo.service.loopingcall [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 616.754106] env[61906]: DEBUG nova.compute.manager [-] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 616.754106] env[61906]: DEBUG nova.network.neutron [-] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 616.788079] env[61906]: DEBUG nova.network.neutron [-] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 617.022700] env[61906]: DEBUG nova.network.neutron [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Successfully created port: 8b413ae7-89f7-491e-bb59-a9fd1b6547af {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 617.140748] env[61906]: DEBUG nova.compute.manager [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 617.182690] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Releasing lock "refresh_cache-eb4cac2e-85d8-46b6-a4d4-ff12ee70e435" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.186352] env[61906]: DEBUG nova.compute.manager [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 617.186352] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 617.186352] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b17bc2-3604-4705-9885-ad956418a6be {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.196221] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 617.196466] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5afa688c-5290-48b7-85ed-bb4323b3811b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.202959] env[61906]: DEBUG oslo_vmware.api [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Waiting for the task: (returnval){ [ 617.202959] env[61906]: value = "task-1333095" [ 617.202959] env[61906]: _type = "Task" [ 617.202959] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.215875] env[61906]: DEBUG oslo_vmware.api [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333095, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.293963] env[61906]: DEBUG nova.network.neutron [-] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.322431] env[61906]: DEBUG nova.compute.manager [req-2a14f2f0-542b-4dfc-8f70-cd51b9b27044 req-ec8ed634-1da4-4587-a9da-af9d5a7970d4 service nova] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Received event network-changed-0206d2c8-6ff4-4df4-a6e2-b6a3d7e4c46a {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 617.322577] env[61906]: DEBUG nova.compute.manager [req-2a14f2f0-542b-4dfc-8f70-cd51b9b27044 req-ec8ed634-1da4-4587-a9da-af9d5a7970d4 service nova] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Refreshing instance network info cache due to event network-changed-0206d2c8-6ff4-4df4-a6e2-b6a3d7e4c46a. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 617.322838] env[61906]: DEBUG oslo_concurrency.lockutils [req-2a14f2f0-542b-4dfc-8f70-cd51b9b27044 req-ec8ed634-1da4-4587-a9da-af9d5a7970d4 service nova] Acquiring lock "refresh_cache-bff5e099-f6ad-41bd-a55a-6987310c37cf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.323809] env[61906]: DEBUG oslo_concurrency.lockutils [req-2a14f2f0-542b-4dfc-8f70-cd51b9b27044 req-ec8ed634-1da4-4587-a9da-af9d5a7970d4 service nova] Acquired lock "refresh_cache-bff5e099-f6ad-41bd-a55a-6987310c37cf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.323809] env[61906]: DEBUG nova.network.neutron [req-2a14f2f0-542b-4dfc-8f70-cd51b9b27044 req-ec8ed634-1da4-4587-a9da-af9d5a7970d4 service nova] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Refreshing network info cache for port 0206d2c8-6ff4-4df4-a6e2-b6a3d7e4c46a {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 617.431499] env[61906]: DEBUG nova.virt.hardware [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 617.431789] env[61906]: DEBUG nova.virt.hardware [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 617.431973] env[61906]: DEBUG nova.virt.hardware [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 617.432248] env[61906]: DEBUG nova.virt.hardware [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 617.432456] env[61906]: DEBUG nova.virt.hardware [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 617.432646] env[61906]: DEBUG nova.virt.hardware [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 617.432892] env[61906]: DEBUG nova.virt.hardware [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 617.433100] env[61906]: DEBUG nova.virt.hardware [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 617.433792] env[61906]: DEBUG nova.virt.hardware [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 617.433792] env[61906]: DEBUG nova.virt.hardware [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 617.433792] env[61906]: DEBUG nova.virt.hardware [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 617.435028] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fec987be-b80c-43ae-8d3e-68330a9c4def {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.444970] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ccf79b-c154-4cd6-89be-1e989550aa9d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.459301] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Instance VIF info [] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 617.465443] env[61906]: DEBUG oslo.service.loopingcall [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 617.468588] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 617.472016] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8c49f09a-ff3d-4121-9a31-54ac4d34036e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.493188] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 617.493188] env[61906]: value = "task-1333096" [ 617.493188] env[61906]: _type = "Task" [ 617.493188] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.502340] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333096, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.714563] env[61906]: DEBUG oslo_vmware.api [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333095, 'name': PowerOffVM_Task, 'duration_secs': 0.177246} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.716204] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 617.716680] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 617.717218] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45a56fc1-5960-4333-8dbe-5af53a394b54 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.720247] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e36ae6bd-a796-43cb-b75c-55009efaab7a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.726140] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d00f3a88-a35a-4067-b71c-224ad9d12a0a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.757988] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2211bfca-e75f-42ab-ac08-0bcd2c802559 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.760729] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 617.760951] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 617.761134] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Deleting the datastore file [datastore2] eb4cac2e-85d8-46b6-a4d4-ff12ee70e435 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 617.761411] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a228fb15-3564-4946-a88a-d9010bdb9136 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.769122] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df6a40dd-aeae-4811-8014-bb44451dc3a7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.773039] env[61906]: DEBUG oslo_vmware.api [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Waiting for the task: (returnval){ [ 617.773039] env[61906]: value = "task-1333098" [ 617.773039] env[61906]: _type = "Task" [ 617.773039] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.784145] env[61906]: DEBUG nova.compute.provider_tree [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 617.790411] env[61906]: DEBUG oslo_vmware.api [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333098, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.799030] env[61906]: INFO nova.compute.manager [-] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Took 1.05 seconds to deallocate network for instance. [ 617.801158] env[61906]: DEBUG nova.compute.claims [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 617.801394] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.845652] env[61906]: DEBUG nova.network.neutron [req-2a14f2f0-542b-4dfc-8f70-cd51b9b27044 req-ec8ed634-1da4-4587-a9da-af9d5a7970d4 service nova] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 617.910570] env[61906]: DEBUG nova.network.neutron [req-2a14f2f0-542b-4dfc-8f70-cd51b9b27044 req-ec8ed634-1da4-4587-a9da-af9d5a7970d4 service nova] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.003411] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333096, 'name': CreateVM_Task, 'duration_secs': 0.258964} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.003579] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 618.003975] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.004148] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.004469] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 618.004715] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db158831-3a21-47a1-9025-c85f7413c6ef {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.009091] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for the task: (returnval){ [ 618.009091] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5264d0bb-c54e-3ad6-80de-5f0d22076282" [ 618.009091] env[61906]: _type = "Task" [ 618.009091] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.016776] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5264d0bb-c54e-3ad6-80de-5f0d22076282, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.033161] env[61906]: DEBUG nova.compute.manager [req-44b84aad-bd9a-4cc6-80c8-ae41b5ff53db req-9bcc0014-04c7-4274-b544-00c075f25c4c service nova] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Received event network-changed-8b413ae7-89f7-491e-bb59-a9fd1b6547af {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 618.033406] env[61906]: DEBUG nova.compute.manager [req-44b84aad-bd9a-4cc6-80c8-ae41b5ff53db req-9bcc0014-04c7-4274-b544-00c075f25c4c service nova] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Refreshing instance network info cache due to event network-changed-8b413ae7-89f7-491e-bb59-a9fd1b6547af. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 618.033627] env[61906]: DEBUG oslo_concurrency.lockutils [req-44b84aad-bd9a-4cc6-80c8-ae41b5ff53db req-9bcc0014-04c7-4274-b544-00c075f25c4c service nova] Acquiring lock "refresh_cache-e64791e3-7269-4953-bfe1-fe776744f493" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.033800] env[61906]: DEBUG oslo_concurrency.lockutils [req-44b84aad-bd9a-4cc6-80c8-ae41b5ff53db req-9bcc0014-04c7-4274-b544-00c075f25c4c service nova] Acquired lock "refresh_cache-e64791e3-7269-4953-bfe1-fe776744f493" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.033994] env[61906]: DEBUG nova.network.neutron [req-44b84aad-bd9a-4cc6-80c8-ae41b5ff53db req-9bcc0014-04c7-4274-b544-00c075f25c4c service nova] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Refreshing network info cache for port 8b413ae7-89f7-491e-bb59-a9fd1b6547af {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 618.159942] env[61906]: DEBUG nova.compute.manager [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 618.162962] env[61906]: ERROR nova.compute.manager [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8b413ae7-89f7-491e-bb59-a9fd1b6547af, please check neutron logs for more information. [ 618.162962] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 618.162962] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.162962] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 618.162962] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.162962] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 618.162962] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.162962] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 618.162962] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.162962] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 618.162962] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.162962] env[61906]: ERROR nova.compute.manager raise self.value [ 618.162962] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.162962] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 618.162962] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.162962] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 618.164022] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.164022] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 618.164022] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8b413ae7-89f7-491e-bb59-a9fd1b6547af, please check neutron logs for more information. [ 618.164022] env[61906]: ERROR nova.compute.manager [ 618.164022] env[61906]: Traceback (most recent call last): [ 618.164022] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 618.164022] env[61906]: listener.cb(fileno) [ 618.164022] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.164022] env[61906]: result = function(*args, **kwargs) [ 618.164022] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 618.164022] env[61906]: return func(*args, **kwargs) [ 618.164022] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.164022] env[61906]: raise e [ 618.164022] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.164022] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 618.164022] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.164022] env[61906]: created_port_ids = self._update_ports_for_instance( [ 618.164022] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.164022] env[61906]: with excutils.save_and_reraise_exception(): [ 618.164022] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.164022] env[61906]: self.force_reraise() [ 618.164022] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.164022] env[61906]: raise self.value [ 618.164022] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.164022] env[61906]: updated_port = self._update_port( [ 618.164022] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.164022] env[61906]: _ensure_no_port_binding_failure(port) [ 618.164022] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.164022] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 618.165755] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 8b413ae7-89f7-491e-bb59-a9fd1b6547af, please check neutron logs for more information. [ 618.165755] env[61906]: Removing descriptor: 18 [ 618.183781] env[61906]: DEBUG nova.virt.hardware [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 618.183781] env[61906]: DEBUG nova.virt.hardware [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 618.183964] env[61906]: DEBUG nova.virt.hardware [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 618.184158] env[61906]: DEBUG nova.virt.hardware [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 618.184307] env[61906]: DEBUG nova.virt.hardware [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 618.184453] env[61906]: DEBUG nova.virt.hardware [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 618.184650] env[61906]: DEBUG nova.virt.hardware [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 618.184802] env[61906]: DEBUG nova.virt.hardware [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 618.184962] env[61906]: DEBUG nova.virt.hardware [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 618.185178] env[61906]: DEBUG nova.virt.hardware [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 618.185374] env[61906]: DEBUG nova.virt.hardware [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 618.186477] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-968d6515-3eb5-4f71-8408-507f85e07371 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.194475] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efcca37b-8877-4de1-a125-89b0a7855c50 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.208064] env[61906]: ERROR nova.compute.manager [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8b413ae7-89f7-491e-bb59-a9fd1b6547af, please check neutron logs for more information. [ 618.208064] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] Traceback (most recent call last): [ 618.208064] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 618.208064] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] yield resources [ 618.208064] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 618.208064] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] self.driver.spawn(context, instance, image_meta, [ 618.208064] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 618.208064] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] self._vmops.spawn(context, instance, image_meta, injected_files, [ 618.208064] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 618.208064] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] vm_ref = self.build_virtual_machine(instance, [ 618.208064] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.208476] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.208476] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.208476] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] for vif in network_info: [ 618.208476] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 618.208476] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] return self._sync_wrapper(fn, *args, **kwargs) [ 618.208476] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 618.208476] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] self.wait() [ 618.208476] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 618.208476] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] self[:] = self._gt.wait() [ 618.208476] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.208476] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] return self._exit_event.wait() [ 618.208476] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 618.208476] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] current.throw(*self._exc) [ 618.209044] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.209044] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] result = function(*args, **kwargs) [ 618.209044] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 618.209044] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] return func(*args, **kwargs) [ 618.209044] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.209044] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] raise e [ 618.209044] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.209044] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] nwinfo = self.network_api.allocate_for_instance( [ 618.209044] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.209044] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] created_port_ids = self._update_ports_for_instance( [ 618.209044] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.209044] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] with excutils.save_and_reraise_exception(): [ 618.209044] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.209474] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] self.force_reraise() [ 618.209474] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.209474] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] raise self.value [ 618.209474] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.209474] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] updated_port = self._update_port( [ 618.209474] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.209474] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] _ensure_no_port_binding_failure(port) [ 618.209474] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.209474] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] raise exception.PortBindingFailed(port_id=port['id']) [ 618.209474] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] nova.exception.PortBindingFailed: Binding failed for port 8b413ae7-89f7-491e-bb59-a9fd1b6547af, please check neutron logs for more information. [ 618.209474] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] [ 618.209474] env[61906]: INFO nova.compute.manager [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Terminating instance [ 618.210481] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Acquiring lock "refresh_cache-e64791e3-7269-4953-bfe1-fe776744f493" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.282888] env[61906]: DEBUG oslo_vmware.api [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Task: {'id': task-1333098, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.102224} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.282888] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 618.282888] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 618.282888] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 618.282888] env[61906]: INFO nova.compute.manager [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Took 1.10 seconds to destroy the instance on the hypervisor. [ 618.283206] env[61906]: DEBUG oslo.service.loopingcall [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 618.283206] env[61906]: DEBUG nova.compute.manager [-] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 618.283310] env[61906]: DEBUG nova.network.neutron [-] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 618.287742] env[61906]: DEBUG nova.scheduler.client.report [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 618.297540] env[61906]: DEBUG nova.network.neutron [-] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.413327] env[61906]: DEBUG oslo_concurrency.lockutils [req-2a14f2f0-542b-4dfc-8f70-cd51b9b27044 req-ec8ed634-1da4-4587-a9da-af9d5a7970d4 service nova] Releasing lock "refresh_cache-bff5e099-f6ad-41bd-a55a-6987310c37cf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.413568] env[61906]: DEBUG nova.compute.manager [req-2a14f2f0-542b-4dfc-8f70-cd51b9b27044 req-ec8ed634-1da4-4587-a9da-af9d5a7970d4 service nova] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Received event network-vif-deleted-0206d2c8-6ff4-4df4-a6e2-b6a3d7e4c46a {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 618.520186] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5264d0bb-c54e-3ad6-80de-5f0d22076282, 'name': SearchDatastore_Task, 'duration_secs': 0.009175} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.520501] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.520730] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 618.520962] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.521125] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.521308] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 618.521563] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e582fbc1-0666-439b-8dbc-8e243cde4e14 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.530356] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 618.530531] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 618.531415] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f84b07a-f3fa-4823-a2e3-d285e53e536a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.538689] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for the task: (returnval){ [ 618.538689] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]521fd472-1b26-7b22-ca20-52bcdb6950dc" [ 618.538689] env[61906]: _type = "Task" [ 618.538689] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.547730] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]521fd472-1b26-7b22-ca20-52bcdb6950dc, 'name': SearchDatastore_Task, 'duration_secs': 0.007376} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.548489] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33998144-5b1b-4866-9c8b-071582a92f39 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.553456] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for the task: (returnval){ [ 618.553456] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52b9dab3-62a5-31b7-e94a-b3ed87a2256e" [ 618.553456] env[61906]: _type = "Task" [ 618.553456] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.558472] env[61906]: DEBUG nova.network.neutron [req-44b84aad-bd9a-4cc6-80c8-ae41b5ff53db req-9bcc0014-04c7-4274-b544-00c075f25c4c service nova] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.563375] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52b9dab3-62a5-31b7-e94a-b3ed87a2256e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.652475] env[61906]: DEBUG nova.network.neutron [req-44b84aad-bd9a-4cc6-80c8-ae41b5ff53db req-9bcc0014-04c7-4274-b544-00c075f25c4c service nova] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.792410] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.663s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.792936] env[61906]: DEBUG nova.compute.manager [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 618.798169] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.847s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.799551] env[61906]: DEBUG nova.network.neutron [-] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.069199] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52b9dab3-62a5-31b7-e94a-b3ed87a2256e, 'name': SearchDatastore_Task, 'duration_secs': 0.008501} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.070093] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.070486] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 318a3224-a156-4354-a73b-460168c0eb9c/318a3224-a156-4354-a73b-460168c0eb9c.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 619.070920] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2cd3a6f4-ea2d-4593-b944-a452f4115165 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.078255] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for the task: (returnval){ [ 619.078255] env[61906]: value = "task-1333099" [ 619.078255] env[61906]: _type = "Task" [ 619.078255] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.086337] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333099, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.155720] env[61906]: DEBUG oslo_concurrency.lockutils [req-44b84aad-bd9a-4cc6-80c8-ae41b5ff53db req-9bcc0014-04c7-4274-b544-00c075f25c4c service nova] Releasing lock "refresh_cache-e64791e3-7269-4953-bfe1-fe776744f493" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.156210] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Acquired lock "refresh_cache-e64791e3-7269-4953-bfe1-fe776744f493" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.156426] env[61906]: DEBUG nova.network.neutron [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 619.304187] env[61906]: DEBUG nova.compute.utils [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 619.309954] env[61906]: DEBUG nova.compute.manager [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 619.310403] env[61906]: DEBUG nova.network.neutron [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 619.312877] env[61906]: INFO nova.compute.manager [-] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Took 1.03 seconds to deallocate network for instance. [ 619.585834] env[61906]: DEBUG nova.policy [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '133c4e406bc54735b60efffc17df8d9a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0128365a1d15449cb778cf8c6d0ae6c1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 619.592629] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333099, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.455606} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.592883] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 318a3224-a156-4354-a73b-460168c0eb9c/318a3224-a156-4354-a73b-460168c0eb9c.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 619.593110] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 619.593349] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e75a8a89-e8a9-4fe4-92b2-4394c429b3ae {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.599768] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for the task: (returnval){ [ 619.599768] env[61906]: value = "task-1333100" [ 619.599768] env[61906]: _type = "Task" [ 619.599768] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.610762] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333100, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.679120] env[61906]: DEBUG nova.network.neutron [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.785871] env[61906]: DEBUG nova.network.neutron [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.807433] env[61906]: DEBUG nova.compute.manager [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 619.819488] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.882656] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8084d4f-080f-4b3b-9ecb-a9b1f800f074 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.890296] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65c1959f-1171-4bba-9993-784d86d803a1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.920996] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d511fe3b-bd8d-4de0-a691-2cf5edf89f59 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.928048] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25c7ef7a-49c4-42da-bba6-8862c342943a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.940880] env[61906]: DEBUG nova.compute.provider_tree [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.021462] env[61906]: DEBUG nova.network.neutron [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Successfully created port: 4c761c95-c60e-4705-ab93-a0e344175aff {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 620.088680] env[61906]: DEBUG nova.compute.manager [req-fc62d576-3935-402f-875f-f533ed6e7022 req-6d4bd35f-08d7-40be-a018-412afbf734b3 service nova] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Received event network-vif-deleted-8b413ae7-89f7-491e-bb59-a9fd1b6547af {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 620.108696] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333100, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064096} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.108961] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 620.109826] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef5fa8b5-1d70-48be-b626-a387da636cb9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.128980] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Reconfiguring VM instance instance-0000000d to attach disk [datastore2] 318a3224-a156-4354-a73b-460168c0eb9c/318a3224-a156-4354-a73b-460168c0eb9c.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 620.129640] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ed7c775-b07e-4f18-844d-053ea7e087a2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.148831] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for the task: (returnval){ [ 620.148831] env[61906]: value = "task-1333101" [ 620.148831] env[61906]: _type = "Task" [ 620.148831] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.162106] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333101, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.288301] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Releasing lock "refresh_cache-e64791e3-7269-4953-bfe1-fe776744f493" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.289862] env[61906]: DEBUG nova.compute.manager [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 620.290128] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 620.290432] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-244ec621-7f83-4437-9497-afc874e12e33 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.300534] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-340b6269-d4bd-48f8-a53e-0f6550ea1624 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.324797] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e64791e3-7269-4953-bfe1-fe776744f493 could not be found. [ 620.325072] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 620.325262] env[61906]: INFO nova.compute.manager [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Took 0.04 seconds to destroy the instance on the hypervisor. [ 620.325702] env[61906]: DEBUG oslo.service.loopingcall [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 620.326010] env[61906]: DEBUG nova.compute.manager [-] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 620.326052] env[61906]: DEBUG nova.network.neutron [-] [instance: e64791e3-7269-4953-bfe1-fe776744f493] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 620.342620] env[61906]: DEBUG nova.network.neutron [-] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.444431] env[61906]: DEBUG nova.scheduler.client.report [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 620.658922] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333101, 'name': ReconfigVM_Task, 'duration_secs': 0.302601} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.659239] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Reconfigured VM instance instance-0000000d to attach disk [datastore2] 318a3224-a156-4354-a73b-460168c0eb9c/318a3224-a156-4354-a73b-460168c0eb9c.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 620.659845] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-09204b3e-d888-4f8e-8883-639eae351f62 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.666596] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for the task: (returnval){ [ 620.666596] env[61906]: value = "task-1333102" [ 620.666596] env[61906]: _type = "Task" [ 620.666596] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.676287] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333102, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.819799] env[61906]: DEBUG nova.compute.manager [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 620.844305] env[61906]: DEBUG nova.virt.hardware [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 620.844548] env[61906]: DEBUG nova.virt.hardware [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 620.844701] env[61906]: DEBUG nova.virt.hardware [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 620.844878] env[61906]: DEBUG nova.virt.hardware [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 620.845034] env[61906]: DEBUG nova.virt.hardware [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 620.845186] env[61906]: DEBUG nova.virt.hardware [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 620.845390] env[61906]: DEBUG nova.virt.hardware [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 620.845548] env[61906]: DEBUG nova.virt.hardware [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 620.845711] env[61906]: DEBUG nova.virt.hardware [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 620.845870] env[61906]: DEBUG nova.virt.hardware [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 620.846060] env[61906]: DEBUG nova.virt.hardware [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 620.846375] env[61906]: DEBUG nova.network.neutron [-] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.847979] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f0bba5-e896-439c-b5a9-c50a4eaa5faa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.857512] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a698663-94ca-43e7-8719-b63998133a10 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.949579] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.153s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.950400] env[61906]: ERROR nova.compute.manager [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e59486b6-bb9d-4623-858d-2956318d9053, please check neutron logs for more information. [ 620.950400] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] Traceback (most recent call last): [ 620.950400] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 620.950400] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] self.driver.spawn(context, instance, image_meta, [ 620.950400] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 620.950400] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] self._vmops.spawn(context, instance, image_meta, injected_files, [ 620.950400] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 620.950400] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] vm_ref = self.build_virtual_machine(instance, [ 620.950400] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 620.950400] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] vif_infos = vmwarevif.get_vif_info(self._session, [ 620.950400] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 620.950774] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] for vif in network_info: [ 620.950774] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 620.950774] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] return self._sync_wrapper(fn, *args, **kwargs) [ 620.950774] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 620.950774] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] self.wait() [ 620.950774] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 620.950774] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] self[:] = self._gt.wait() [ 620.950774] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 620.950774] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] return self._exit_event.wait() [ 620.950774] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 620.950774] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] result = hub.switch() [ 620.950774] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 620.950774] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] return self.greenlet.switch() [ 620.951196] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 620.951196] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] result = function(*args, **kwargs) [ 620.951196] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 620.951196] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] return func(*args, **kwargs) [ 620.951196] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 620.951196] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] raise e [ 620.951196] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 620.951196] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] nwinfo = self.network_api.allocate_for_instance( [ 620.951196] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.951196] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] created_port_ids = self._update_ports_for_instance( [ 620.951196] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.951196] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] with excutils.save_and_reraise_exception(): [ 620.951196] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.951585] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] self.force_reraise() [ 620.951585] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.951585] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] raise self.value [ 620.951585] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.951585] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] updated_port = self._update_port( [ 620.951585] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.951585] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] _ensure_no_port_binding_failure(port) [ 620.951585] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.951585] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] raise exception.PortBindingFailed(port_id=port['id']) [ 620.951585] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] nova.exception.PortBindingFailed: Binding failed for port e59486b6-bb9d-4623-858d-2956318d9053, please check neutron logs for more information. [ 620.951585] env[61906]: ERROR nova.compute.manager [instance: 263da5ea-892b-4142-b487-96444abc97be] [ 620.951904] env[61906]: DEBUG nova.compute.utils [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Binding failed for port e59486b6-bb9d-4623-858d-2956318d9053, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 620.952511] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.403s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.955556] env[61906]: DEBUG nova.compute.manager [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Build of instance 263da5ea-892b-4142-b487-96444abc97be was re-scheduled: Binding failed for port e59486b6-bb9d-4623-858d-2956318d9053, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 620.956018] env[61906]: DEBUG nova.compute.manager [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 620.956277] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Acquiring lock "refresh_cache-263da5ea-892b-4142-b487-96444abc97be" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.956457] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Acquired lock "refresh_cache-263da5ea-892b-4142-b487-96444abc97be" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.956646] env[61906]: DEBUG nova.network.neutron [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 621.059029] env[61906]: ERROR nova.compute.manager [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4c761c95-c60e-4705-ab93-a0e344175aff, please check neutron logs for more information. [ 621.059029] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 621.059029] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.059029] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 621.059029] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.059029] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 621.059029] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.059029] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 621.059029] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.059029] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 621.059029] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.059029] env[61906]: ERROR nova.compute.manager raise self.value [ 621.059029] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.059029] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 621.059029] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.059029] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 621.059672] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.059672] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 621.059672] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4c761c95-c60e-4705-ab93-a0e344175aff, please check neutron logs for more information. [ 621.059672] env[61906]: ERROR nova.compute.manager [ 621.059672] env[61906]: Traceback (most recent call last): [ 621.059672] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 621.059672] env[61906]: listener.cb(fileno) [ 621.059672] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.059672] env[61906]: result = function(*args, **kwargs) [ 621.059672] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 621.059672] env[61906]: return func(*args, **kwargs) [ 621.059672] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.059672] env[61906]: raise e [ 621.059672] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.059672] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 621.059672] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.059672] env[61906]: created_port_ids = self._update_ports_for_instance( [ 621.059672] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.059672] env[61906]: with excutils.save_and_reraise_exception(): [ 621.059672] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.059672] env[61906]: self.force_reraise() [ 621.059672] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.059672] env[61906]: raise self.value [ 621.059672] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.059672] env[61906]: updated_port = self._update_port( [ 621.059672] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.059672] env[61906]: _ensure_no_port_binding_failure(port) [ 621.059672] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.059672] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 621.060785] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 4c761c95-c60e-4705-ab93-a0e344175aff, please check neutron logs for more information. [ 621.060785] env[61906]: Removing descriptor: 18 [ 621.060785] env[61906]: ERROR nova.compute.manager [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4c761c95-c60e-4705-ab93-a0e344175aff, please check neutron logs for more information. [ 621.060785] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Traceback (most recent call last): [ 621.060785] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 621.060785] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] yield resources [ 621.060785] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 621.060785] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] self.driver.spawn(context, instance, image_meta, [ 621.060785] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 621.060785] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] self._vmops.spawn(context, instance, image_meta, injected_files, [ 621.060785] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 621.060785] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] vm_ref = self.build_virtual_machine(instance, [ 621.061217] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 621.061217] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] vif_infos = vmwarevif.get_vif_info(self._session, [ 621.061217] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 621.061217] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] for vif in network_info: [ 621.061217] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 621.061217] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] return self._sync_wrapper(fn, *args, **kwargs) [ 621.061217] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 621.061217] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] self.wait() [ 621.061217] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 621.061217] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] self[:] = self._gt.wait() [ 621.061217] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 621.061217] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] return self._exit_event.wait() [ 621.061217] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 621.061660] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] result = hub.switch() [ 621.061660] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 621.061660] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] return self.greenlet.switch() [ 621.061660] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.061660] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] result = function(*args, **kwargs) [ 621.061660] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 621.061660] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] return func(*args, **kwargs) [ 621.061660] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.061660] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] raise e [ 621.061660] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.061660] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] nwinfo = self.network_api.allocate_for_instance( [ 621.061660] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.061660] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] created_port_ids = self._update_ports_for_instance( [ 621.062105] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.062105] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] with excutils.save_and_reraise_exception(): [ 621.062105] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.062105] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] self.force_reraise() [ 621.062105] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.062105] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] raise self.value [ 621.062105] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.062105] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] updated_port = self._update_port( [ 621.062105] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.062105] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] _ensure_no_port_binding_failure(port) [ 621.062105] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.062105] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] raise exception.PortBindingFailed(port_id=port['id']) [ 621.062519] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] nova.exception.PortBindingFailed: Binding failed for port 4c761c95-c60e-4705-ab93-a0e344175aff, please check neutron logs for more information. [ 621.062519] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] [ 621.062519] env[61906]: INFO nova.compute.manager [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Terminating instance [ 621.062631] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Acquiring lock "refresh_cache-9fe29bcd-a682-4973-9bb8-069adc204086" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.062896] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Acquired lock "refresh_cache-9fe29bcd-a682-4973-9bb8-069adc204086" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.062896] env[61906]: DEBUG nova.network.neutron [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 621.177585] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333102, 'name': Rename_Task, 'duration_secs': 0.141449} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.177859] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 621.178095] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ccbd226c-eb4d-4602-80dd-ea7127329c82 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.184788] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for the task: (returnval){ [ 621.184788] env[61906]: value = "task-1333103" [ 621.184788] env[61906]: _type = "Task" [ 621.184788] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.192140] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333103, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.352039] env[61906]: INFO nova.compute.manager [-] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Took 1.03 seconds to deallocate network for instance. [ 621.354527] env[61906]: DEBUG nova.compute.claims [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 621.354704] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.476774] env[61906]: DEBUG nova.network.neutron [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.567779] env[61906]: DEBUG nova.network.neutron [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.582330] env[61906]: DEBUG nova.network.neutron [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.637814] env[61906]: DEBUG nova.network.neutron [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.694315] env[61906]: DEBUG oslo_vmware.api [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333103, 'name': PowerOnVM_Task, 'duration_secs': 0.409557} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.700093] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 621.700093] env[61906]: DEBUG nova.compute.manager [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 621.700093] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0627beb-316e-4484-aaa2-9cd5d261d581 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.957616] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a6444a1-6d20-4d15-ae1c-1d172722a170 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.966254] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96240765-b9ff-49d6-a2e7-ab382f0f5dce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.998509] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fc65446-b679-430e-8b06-7e901a12ed10 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.006451] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4362e0c1-4754-4109-b93f-f73f3d276da0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.020297] env[61906]: DEBUG nova.compute.provider_tree [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 622.070906] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Releasing lock "refresh_cache-263da5ea-892b-4142-b487-96444abc97be" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.071172] env[61906]: DEBUG nova.compute.manager [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 622.071359] env[61906]: DEBUG nova.compute.manager [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 622.071527] env[61906]: DEBUG nova.network.neutron [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 622.087537] env[61906]: DEBUG nova.network.neutron [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.116215] env[61906]: DEBUG nova.compute.manager [req-a465fad7-1f9a-448d-bc46-9673e39a8c94 req-5d2049ea-5685-48e9-9a48-eb966f925004 service nova] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Received event network-changed-4c761c95-c60e-4705-ab93-a0e344175aff {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 622.116569] env[61906]: DEBUG nova.compute.manager [req-a465fad7-1f9a-448d-bc46-9673e39a8c94 req-5d2049ea-5685-48e9-9a48-eb966f925004 service nova] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Refreshing instance network info cache due to event network-changed-4c761c95-c60e-4705-ab93-a0e344175aff. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 622.116765] env[61906]: DEBUG oslo_concurrency.lockutils [req-a465fad7-1f9a-448d-bc46-9673e39a8c94 req-5d2049ea-5685-48e9-9a48-eb966f925004 service nova] Acquiring lock "refresh_cache-9fe29bcd-a682-4973-9bb8-069adc204086" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.139924] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Releasing lock "refresh_cache-9fe29bcd-a682-4973-9bb8-069adc204086" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.140372] env[61906]: DEBUG nova.compute.manager [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 622.140559] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 622.140909] env[61906]: DEBUG oslo_concurrency.lockutils [req-a465fad7-1f9a-448d-bc46-9673e39a8c94 req-5d2049ea-5685-48e9-9a48-eb966f925004 service nova] Acquired lock "refresh_cache-9fe29bcd-a682-4973-9bb8-069adc204086" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.141118] env[61906]: DEBUG nova.network.neutron [req-a465fad7-1f9a-448d-bc46-9673e39a8c94 req-5d2049ea-5685-48e9-9a48-eb966f925004 service nova] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Refreshing network info cache for port 4c761c95-c60e-4705-ab93-a0e344175aff {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 622.142137] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d6431f1f-dbd7-4cd2-bd72-0a96d94fcc1d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.152031] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b48ea23-1728-4ba8-8260-2b52b31002df {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.174389] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9fe29bcd-a682-4973-9bb8-069adc204086 could not be found. [ 622.174634] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 622.174811] env[61906]: INFO nova.compute.manager [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Took 0.03 seconds to destroy the instance on the hypervisor. [ 622.175067] env[61906]: DEBUG oslo.service.loopingcall [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 622.175265] env[61906]: DEBUG nova.compute.manager [-] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 622.175369] env[61906]: DEBUG nova.network.neutron [-] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 622.215622] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.318813] env[61906]: DEBUG nova.network.neutron [-] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.523367] env[61906]: DEBUG nova.scheduler.client.report [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 622.589517] env[61906]: DEBUG nova.network.neutron [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.681587] env[61906]: DEBUG nova.network.neutron [req-a465fad7-1f9a-448d-bc46-9673e39a8c94 req-5d2049ea-5685-48e9-9a48-eb966f925004 service nova] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.730868] env[61906]: DEBUG nova.network.neutron [req-a465fad7-1f9a-448d-bc46-9673e39a8c94 req-5d2049ea-5685-48e9-9a48-eb966f925004 service nova] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.821566] env[61906]: DEBUG nova.network.neutron [-] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.028700] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.076s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.029975] env[61906]: ERROR nova.compute.manager [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d43156d0-1126-4b2d-8862-1fbd2a120665, please check neutron logs for more information. [ 623.029975] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Traceback (most recent call last): [ 623.029975] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 623.029975] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] self.driver.spawn(context, instance, image_meta, [ 623.029975] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 623.029975] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] self._vmops.spawn(context, instance, image_meta, injected_files, [ 623.029975] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 623.029975] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] vm_ref = self.build_virtual_machine(instance, [ 623.029975] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 623.029975] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] vif_infos = vmwarevif.get_vif_info(self._session, [ 623.029975] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 623.030508] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] for vif in network_info: [ 623.030508] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 623.030508] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] return self._sync_wrapper(fn, *args, **kwargs) [ 623.030508] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 623.030508] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] self.wait() [ 623.030508] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 623.030508] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] self[:] = self._gt.wait() [ 623.030508] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 623.030508] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] return self._exit_event.wait() [ 623.030508] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 623.030508] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] result = hub.switch() [ 623.030508] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 623.030508] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] return self.greenlet.switch() [ 623.031196] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.031196] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] result = function(*args, **kwargs) [ 623.031196] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 623.031196] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] return func(*args, **kwargs) [ 623.031196] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 623.031196] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] raise e [ 623.031196] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.031196] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] nwinfo = self.network_api.allocate_for_instance( [ 623.031196] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 623.031196] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] created_port_ids = self._update_ports_for_instance( [ 623.031196] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 623.031196] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] with excutils.save_and_reraise_exception(): [ 623.031196] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.031594] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] self.force_reraise() [ 623.031594] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.031594] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] raise self.value [ 623.031594] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 623.031594] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] updated_port = self._update_port( [ 623.031594] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.031594] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] _ensure_no_port_binding_failure(port) [ 623.031594] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.031594] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] raise exception.PortBindingFailed(port_id=port['id']) [ 623.031594] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] nova.exception.PortBindingFailed: Binding failed for port d43156d0-1126-4b2d-8862-1fbd2a120665, please check neutron logs for more information. [ 623.031594] env[61906]: ERROR nova.compute.manager [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] [ 623.031964] env[61906]: DEBUG nova.compute.utils [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Binding failed for port d43156d0-1126-4b2d-8862-1fbd2a120665, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 623.031964] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.519s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.032737] env[61906]: INFO nova.compute.claims [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 623.035349] env[61906]: DEBUG nova.compute.manager [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Build of instance 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121 was re-scheduled: Binding failed for port d43156d0-1126-4b2d-8862-1fbd2a120665, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 623.036072] env[61906]: DEBUG nova.compute.manager [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 623.036072] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Acquiring lock "refresh_cache-1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.036200] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Acquired lock "refresh_cache-1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.036280] env[61906]: DEBUG nova.network.neutron [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 623.092795] env[61906]: INFO nova.compute.manager [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] [instance: 263da5ea-892b-4142-b487-96444abc97be] Took 1.02 seconds to deallocate network for instance. [ 623.132434] env[61906]: INFO nova.compute.manager [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Rebuilding instance [ 623.168593] env[61906]: DEBUG nova.compute.manager [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 623.169445] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e41cf990-3b44-42f5-a0a8-461a5bcc0d5d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.235040] env[61906]: DEBUG oslo_concurrency.lockutils [req-a465fad7-1f9a-448d-bc46-9673e39a8c94 req-5d2049ea-5685-48e9-9a48-eb966f925004 service nova] Releasing lock "refresh_cache-9fe29bcd-a682-4973-9bb8-069adc204086" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.235381] env[61906]: DEBUG nova.compute.manager [req-a465fad7-1f9a-448d-bc46-9673e39a8c94 req-5d2049ea-5685-48e9-9a48-eb966f925004 service nova] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Received event network-vif-deleted-4c761c95-c60e-4705-ab93-a0e344175aff {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 623.323247] env[61906]: INFO nova.compute.manager [-] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Took 1.15 seconds to deallocate network for instance. [ 623.326947] env[61906]: DEBUG nova.compute.claims [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 623.327337] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.563585] env[61906]: DEBUG nova.network.neutron [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 623.622786] env[61906]: DEBUG nova.network.neutron [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.679304] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 623.679592] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e43f2cb9-c151-4880-b7e3-ade21c6cf646 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.687141] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Waiting for the task: (returnval){ [ 623.687141] env[61906]: value = "task-1333104" [ 623.687141] env[61906]: _type = "Task" [ 623.687141] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.696582] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Task: {'id': task-1333104, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.126030] env[61906]: INFO nova.scheduler.client.report [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Deleted allocations for instance 263da5ea-892b-4142-b487-96444abc97be [ 624.136085] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Releasing lock "refresh_cache-1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.136085] env[61906]: DEBUG nova.compute.manager [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 624.136085] env[61906]: DEBUG nova.compute.manager [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 624.136085] env[61906]: DEBUG nova.network.neutron [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 624.154818] env[61906]: DEBUG nova.network.neutron [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.197124] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Task: {'id': task-1333104, 'name': PowerOffVM_Task, 'duration_secs': 0.123705} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.199543] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 624.199761] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 624.200698] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d11fa003-0df2-485f-97e2-47a9aed88320 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.211937] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 624.212302] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c45010b-d9ad-4505-a57d-b67619569ef2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.237417] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 624.237685] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 624.237855] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Deleting the datastore file [datastore2] 318a3224-a156-4354-a73b-460168c0eb9c {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 624.240371] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1b365309-538f-42ba-875d-c81ec16653d7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.246797] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Waiting for the task: (returnval){ [ 624.246797] env[61906]: value = "task-1333106" [ 624.246797] env[61906]: _type = "Task" [ 624.246797] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.254197] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Task: {'id': task-1333106, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.538521] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c698126f-1fd5-41ee-bac3-b4c9bd375099 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.546208] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b984d483-2720-4023-9606-c5bac8de8811 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.575150] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-468db737-1d7e-4cb3-944a-cc38b843ad6b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.582245] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae8b0f5f-6b7d-46a2-9e8d-32d5f796c993 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.594962] env[61906]: DEBUG nova.compute.provider_tree [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.640042] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5ad943a1-5d28-4390-9e7f-a3deeadf0dbf tempest-VolumesAssistedSnapshotsTest-1476113900 tempest-VolumesAssistedSnapshotsTest-1476113900-project-member] Lock "263da5ea-892b-4142-b487-96444abc97be" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 58.684s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.656328] env[61906]: DEBUG nova.network.neutron [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.758405] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Task: {'id': task-1333106, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.110348} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.758405] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 624.758405] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 624.758405] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 625.098444] env[61906]: DEBUG nova.scheduler.client.report [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 625.142982] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 625.161775] env[61906]: INFO nova.compute.manager [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] [instance: 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121] Took 1.03 seconds to deallocate network for instance. [ 625.603025] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.572s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.603650] env[61906]: DEBUG nova.compute.manager [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 625.606202] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.515s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.607701] env[61906]: INFO nova.compute.claims [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 625.671545] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.790692] env[61906]: DEBUG nova.virt.hardware [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 625.790846] env[61906]: DEBUG nova.virt.hardware [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 625.790981] env[61906]: DEBUG nova.virt.hardware [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 625.791178] env[61906]: DEBUG nova.virt.hardware [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 625.791319] env[61906]: DEBUG nova.virt.hardware [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 625.791468] env[61906]: DEBUG nova.virt.hardware [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 625.791671] env[61906]: DEBUG nova.virt.hardware [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 625.791828] env[61906]: DEBUG nova.virt.hardware [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 625.791994] env[61906]: DEBUG nova.virt.hardware [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 625.792182] env[61906]: DEBUG nova.virt.hardware [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 625.792357] env[61906]: DEBUG nova.virt.hardware [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 625.793244] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03f32750-5f37-4afe-931d-885343f16cb1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.801774] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ef79d83-5615-431d-8e0a-a898292a5310 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.814885] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Instance VIF info [] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 625.821030] env[61906]: DEBUG oslo.service.loopingcall [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 625.821135] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 625.821270] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e7fb42cc-81ba-42da-8da7-9ea92746a9d9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.839426] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 625.839426] env[61906]: value = "task-1333107" [ 625.839426] env[61906]: _type = "Task" [ 625.839426] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.849842] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333107, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.111918] env[61906]: DEBUG nova.compute.utils [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 626.115468] env[61906]: DEBUG nova.compute.manager [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 626.116193] env[61906]: DEBUG nova.network.neutron [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 626.177500] env[61906]: DEBUG nova.policy [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '92238df196bb43348494352835ab8f52', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c1e22d7299204152bf28c4effb6f3a3f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 626.193561] env[61906]: INFO nova.scheduler.client.report [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Deleted allocations for instance 1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121 [ 626.352715] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333107, 'name': CreateVM_Task, 'duration_secs': 0.235255} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.352957] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 626.353305] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.353469] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.353792] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 626.354399] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f80ecb04-b7fe-40bb-82f0-a6b032d81cb7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.359345] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Waiting for the task: (returnval){ [ 626.359345] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]525d5df3-4726-8e01-6fdb-7110364eded9" [ 626.359345] env[61906]: _type = "Task" [ 626.359345] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.366995] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]525d5df3-4726-8e01-6fdb-7110364eded9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.519445] env[61906]: DEBUG nova.network.neutron [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Successfully created port: bbd06d6f-2ab5-40f7-ba65-fe54621ee57b {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 626.616234] env[61906]: DEBUG nova.compute.manager [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 626.709017] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46dfac03-ee01-41b0-9896-dd186e458ba0 tempest-ListImageFiltersTestJSON-1110095908 tempest-ListImageFiltersTestJSON-1110095908-project-member] Lock "1fa06c5a-8ecb-4e52-9eba-cbcf63dd9121" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.761s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.874551] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]525d5df3-4726-8e01-6fdb-7110364eded9, 'name': SearchDatastore_Task, 'duration_secs': 0.008353} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.874899] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.875161] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 626.875400] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.875545] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.875723] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 626.875999] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c4fc7f93-871b-46e4-ac1e-1e65888842b5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.900143] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 626.900143] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 626.900143] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d34f35b6-0059-4b7d-aad4-5fb207ef04ca {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.903270] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Waiting for the task: (returnval){ [ 626.903270] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52b96616-1feb-135f-e559-f7f3e9aa18f5" [ 626.903270] env[61906]: _type = "Task" [ 626.903270] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.922117] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52b96616-1feb-135f-e559-f7f3e9aa18f5, 'name': SearchDatastore_Task, 'duration_secs': 0.008344} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.926125] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b952380f-6497-4ce3-9826-f85a0a57523b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.931979] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Waiting for the task: (returnval){ [ 626.931979] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5257b1c4-7069-b191-64b8-26ccf18c86e9" [ 626.931979] env[61906]: _type = "Task" [ 626.931979] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.943500] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5257b1c4-7069-b191-64b8-26ccf18c86e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.183113] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a92120b8-6c77-497d-87b6-5d79ca83869c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.191587] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb63e0d-ade4-4a62-a9a1-1a5178d9376d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.223321] env[61906]: DEBUG nova.compute.manager [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 627.225744] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51117d65-9cc0-4622-9333-bad010c901fe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.233721] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cc0e5ae-9376-4b7c-91f6-9faff6341ef2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.250060] env[61906]: DEBUG nova.compute.provider_tree [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 627.442683] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5257b1c4-7069-b191-64b8-26ccf18c86e9, 'name': SearchDatastore_Task, 'duration_secs': 0.008204} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.442946] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.443216] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 318a3224-a156-4354-a73b-460168c0eb9c/318a3224-a156-4354-a73b-460168c0eb9c.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 627.443474] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7a91acfa-ad2a-4682-8e51-daa466fd893e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.449652] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Waiting for the task: (returnval){ [ 627.449652] env[61906]: value = "task-1333108" [ 627.449652] env[61906]: _type = "Task" [ 627.449652] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.458348] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Task: {'id': task-1333108, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.631823] env[61906]: DEBUG nova.compute.manager [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 627.663312] env[61906]: DEBUG nova.virt.hardware [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 627.663721] env[61906]: DEBUG nova.virt.hardware [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 627.663961] env[61906]: DEBUG nova.virt.hardware [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 627.664234] env[61906]: DEBUG nova.virt.hardware [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 627.664450] env[61906]: DEBUG nova.virt.hardware [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 627.664643] env[61906]: DEBUG nova.virt.hardware [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 627.664905] env[61906]: DEBUG nova.virt.hardware [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 627.665113] env[61906]: DEBUG nova.virt.hardware [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 627.665389] env[61906]: DEBUG nova.virt.hardware [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 627.665566] env[61906]: DEBUG nova.virt.hardware [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 627.665822] env[61906]: DEBUG nova.virt.hardware [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 627.666769] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d86e40d0-34e4-482f-aa42-cfb84d381392 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.679655] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4766c15e-e52e-465e-88c2-768ec5157b69 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.749604] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.752854] env[61906]: DEBUG nova.scheduler.client.report [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 627.966201] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Task: {'id': task-1333108, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.470642} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.966672] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 318a3224-a156-4354-a73b-460168c0eb9c/318a3224-a156-4354-a73b-460168c0eb9c.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 627.967075] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 627.967487] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ef0f3bd0-607c-4f20-b1eb-48adcedd5582 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.974376] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Waiting for the task: (returnval){ [ 627.974376] env[61906]: value = "task-1333109" [ 627.974376] env[61906]: _type = "Task" [ 627.974376] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.982961] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Task: {'id': task-1333109, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.259029] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.653s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.259526] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 628.266082] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.284s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.271380] env[61906]: DEBUG nova.compute.manager [req-c8d4fc08-22d1-4afe-bc49-c940de1cde8b req-0a0ed611-c268-422e-80cf-2878fce5e36c service nova] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Received event network-changed-bbd06d6f-2ab5-40f7-ba65-fe54621ee57b {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 628.271483] env[61906]: DEBUG nova.compute.manager [req-c8d4fc08-22d1-4afe-bc49-c940de1cde8b req-0a0ed611-c268-422e-80cf-2878fce5e36c service nova] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Refreshing instance network info cache due to event network-changed-bbd06d6f-2ab5-40f7-ba65-fe54621ee57b. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 628.271690] env[61906]: DEBUG oslo_concurrency.lockutils [req-c8d4fc08-22d1-4afe-bc49-c940de1cde8b req-0a0ed611-c268-422e-80cf-2878fce5e36c service nova] Acquiring lock "refresh_cache-b4df3af4-f60d-4875-af62-d9c162e4a5c8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.271834] env[61906]: DEBUG oslo_concurrency.lockutils [req-c8d4fc08-22d1-4afe-bc49-c940de1cde8b req-0a0ed611-c268-422e-80cf-2878fce5e36c service nova] Acquired lock "refresh_cache-b4df3af4-f60d-4875-af62-d9c162e4a5c8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.272042] env[61906]: DEBUG nova.network.neutron [req-c8d4fc08-22d1-4afe-bc49-c940de1cde8b req-0a0ed611-c268-422e-80cf-2878fce5e36c service nova] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Refreshing network info cache for port bbd06d6f-2ab5-40f7-ba65-fe54621ee57b {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 628.485567] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Task: {'id': task-1333109, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063161} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.486485] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 628.487458] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb8d5a56-9f37-4336-9716-cc38ad9c40e8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.512127] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Reconfiguring VM instance instance-0000000d to attach disk [datastore2] 318a3224-a156-4354-a73b-460168c0eb9c/318a3224-a156-4354-a73b-460168c0eb9c.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 628.512127] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7b3331d3-eb90-441f-bc68-a502df12a0c1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.534060] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Waiting for the task: (returnval){ [ 628.534060] env[61906]: value = "task-1333110" [ 628.534060] env[61906]: _type = "Task" [ 628.534060] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.543131] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Task: {'id': task-1333110, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.568707] env[61906]: ERROR nova.compute.manager [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bbd06d6f-2ab5-40f7-ba65-fe54621ee57b, please check neutron logs for more information. [ 628.568707] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 628.568707] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.568707] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 628.568707] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 628.568707] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 628.568707] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 628.568707] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 628.568707] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.568707] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 628.568707] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.568707] env[61906]: ERROR nova.compute.manager raise self.value [ 628.568707] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 628.568707] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 628.568707] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.568707] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 628.569272] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.569272] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 628.569272] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bbd06d6f-2ab5-40f7-ba65-fe54621ee57b, please check neutron logs for more information. [ 628.569272] env[61906]: ERROR nova.compute.manager [ 628.569272] env[61906]: Traceback (most recent call last): [ 628.569272] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 628.569272] env[61906]: listener.cb(fileno) [ 628.569272] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.569272] env[61906]: result = function(*args, **kwargs) [ 628.569272] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 628.569272] env[61906]: return func(*args, **kwargs) [ 628.569272] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 628.569272] env[61906]: raise e [ 628.569272] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.569272] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 628.569272] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 628.569272] env[61906]: created_port_ids = self._update_ports_for_instance( [ 628.569272] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 628.569272] env[61906]: with excutils.save_and_reraise_exception(): [ 628.569272] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.569272] env[61906]: self.force_reraise() [ 628.569272] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.569272] env[61906]: raise self.value [ 628.569272] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 628.569272] env[61906]: updated_port = self._update_port( [ 628.569272] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.569272] env[61906]: _ensure_no_port_binding_failure(port) [ 628.569272] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.569272] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 628.570284] env[61906]: nova.exception.PortBindingFailed: Binding failed for port bbd06d6f-2ab5-40f7-ba65-fe54621ee57b, please check neutron logs for more information. [ 628.570284] env[61906]: Removing descriptor: 16 [ 628.570284] env[61906]: ERROR nova.compute.manager [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bbd06d6f-2ab5-40f7-ba65-fe54621ee57b, please check neutron logs for more information. [ 628.570284] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Traceback (most recent call last): [ 628.570284] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 628.570284] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] yield resources [ 628.570284] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 628.570284] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] self.driver.spawn(context, instance, image_meta, [ 628.570284] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 628.570284] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 628.570284] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 628.570284] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] vm_ref = self.build_virtual_machine(instance, [ 628.570738] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 628.570738] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] vif_infos = vmwarevif.get_vif_info(self._session, [ 628.570738] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 628.570738] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] for vif in network_info: [ 628.570738] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 628.570738] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] return self._sync_wrapper(fn, *args, **kwargs) [ 628.570738] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 628.570738] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] self.wait() [ 628.570738] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 628.570738] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] self[:] = self._gt.wait() [ 628.570738] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 628.570738] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] return self._exit_event.wait() [ 628.570738] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 628.572367] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] result = hub.switch() [ 628.572367] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 628.572367] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] return self.greenlet.switch() [ 628.572367] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.572367] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] result = function(*args, **kwargs) [ 628.572367] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 628.572367] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] return func(*args, **kwargs) [ 628.572367] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 628.572367] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] raise e [ 628.572367] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.572367] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] nwinfo = self.network_api.allocate_for_instance( [ 628.572367] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 628.572367] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] created_port_ids = self._update_ports_for_instance( [ 628.572870] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 628.572870] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] with excutils.save_and_reraise_exception(): [ 628.572870] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.572870] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] self.force_reraise() [ 628.572870] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.572870] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] raise self.value [ 628.572870] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 628.572870] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] updated_port = self._update_port( [ 628.572870] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.572870] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] _ensure_no_port_binding_failure(port) [ 628.572870] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.572870] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] raise exception.PortBindingFailed(port_id=port['id']) [ 628.573325] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] nova.exception.PortBindingFailed: Binding failed for port bbd06d6f-2ab5-40f7-ba65-fe54621ee57b, please check neutron logs for more information. [ 628.573325] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] [ 628.573325] env[61906]: INFO nova.compute.manager [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Terminating instance [ 628.573325] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Acquiring lock "refresh_cache-b4df3af4-f60d-4875-af62-d9c162e4a5c8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.771168] env[61906]: DEBUG nova.compute.utils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 628.773159] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 628.773159] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 628.812415] env[61906]: DEBUG nova.network.neutron [req-c8d4fc08-22d1-4afe-bc49-c940de1cde8b req-0a0ed611-c268-422e-80cf-2878fce5e36c service nova] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.851742] env[61906]: DEBUG nova.policy [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d720f18503049d784a1cfe61f242e18', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7a83550c66b84235987277092e630b74', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 628.934211] env[61906]: DEBUG nova.network.neutron [req-c8d4fc08-22d1-4afe-bc49-c940de1cde8b req-0a0ed611-c268-422e-80cf-2878fce5e36c service nova] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.042800] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Task: {'id': task-1333110, 'name': ReconfigVM_Task, 'duration_secs': 0.299347} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.043279] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Reconfigured VM instance instance-0000000d to attach disk [datastore2] 318a3224-a156-4354-a73b-460168c0eb9c/318a3224-a156-4354-a73b-460168c0eb9c.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 629.043791] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ad5ba16c-2fa2-49dd-aba5-c272e1c20674 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.052796] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Waiting for the task: (returnval){ [ 629.052796] env[61906]: value = "task-1333111" [ 629.052796] env[61906]: _type = "Task" [ 629.052796] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.062833] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Task: {'id': task-1333111, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.279963] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 629.286516] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-781ec740-0be5-49d0-823f-7f503db3bad2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.294507] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f5dec93-c960-4b21-a85c-2cbf0f3636cb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.329887] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e703b4ce-adb1-450a-8cec-8eea7016d2a8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.338626] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62d6d822-8658-4b3f-9066-35bc72687956 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.356087] env[61906]: DEBUG nova.compute.provider_tree [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 629.362956] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Successfully created port: 85d0c268-50f0-4953-bc6d-34eba905c8fd {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 629.439395] env[61906]: DEBUG oslo_concurrency.lockutils [req-c8d4fc08-22d1-4afe-bc49-c940de1cde8b req-0a0ed611-c268-422e-80cf-2878fce5e36c service nova] Releasing lock "refresh_cache-b4df3af4-f60d-4875-af62-d9c162e4a5c8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.439894] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Acquired lock "refresh_cache-b4df3af4-f60d-4875-af62-d9c162e4a5c8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.440128] env[61906]: DEBUG nova.network.neutron [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 629.561784] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Task: {'id': task-1333111, 'name': Rename_Task, 'duration_secs': 0.128553} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.562072] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 629.562300] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a94161cf-0e3e-4089-a7dc-679237f92ed3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.568186] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Waiting for the task: (returnval){ [ 629.568186] env[61906]: value = "task-1333112" [ 629.568186] env[61906]: _type = "Task" [ 629.568186] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.575373] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Task: {'id': task-1333112, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.858157] env[61906]: DEBUG nova.scheduler.client.report [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 629.966112] env[61906]: DEBUG nova.network.neutron [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.084340] env[61906]: DEBUG oslo_vmware.api [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Task: {'id': task-1333112, 'name': PowerOnVM_Task, 'duration_secs': 0.4394} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.084621] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 630.085452] env[61906]: DEBUG nova.compute.manager [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 630.086301] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ad484de-8e23-4468-b014-6d3db4ed463c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.125274] env[61906]: DEBUG nova.network.neutron [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.294618] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 630.323591] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 630.324284] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 630.324625] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 630.325039] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 630.325332] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 630.327438] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 630.327438] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 630.327438] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 630.327438] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 630.327438] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 630.327569] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 630.328337] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71640c07-217a-4df1-af30-d1b646ebcc03 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.337721] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04e41fd6-de49-4204-8e7f-53597c9b220e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.364243] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.102s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.364851] env[61906]: ERROR nova.compute.manager [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fa2b5a1c-551c-44cc-bbf8-8f23c6f3e8ab, please check neutron logs for more information. [ 630.364851] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Traceback (most recent call last): [ 630.364851] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 630.364851] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] self.driver.spawn(context, instance, image_meta, [ 630.364851] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 630.364851] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] self._vmops.spawn(context, instance, image_meta, injected_files, [ 630.364851] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 630.364851] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] vm_ref = self.build_virtual_machine(instance, [ 630.364851] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 630.364851] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] vif_infos = vmwarevif.get_vif_info(self._session, [ 630.364851] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 630.365093] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] for vif in network_info: [ 630.365093] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 630.365093] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] return self._sync_wrapper(fn, *args, **kwargs) [ 630.365093] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 630.365093] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] self.wait() [ 630.365093] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 630.365093] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] self[:] = self._gt.wait() [ 630.365093] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 630.365093] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] return self._exit_event.wait() [ 630.365093] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 630.365093] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] result = hub.switch() [ 630.365093] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 630.365093] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] return self.greenlet.switch() [ 630.365422] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.365422] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] result = function(*args, **kwargs) [ 630.365422] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 630.365422] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] return func(*args, **kwargs) [ 630.365422] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 630.365422] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] raise e [ 630.365422] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.365422] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] nwinfo = self.network_api.allocate_for_instance( [ 630.365422] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 630.365422] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] created_port_ids = self._update_ports_for_instance( [ 630.365422] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 630.365422] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] with excutils.save_and_reraise_exception(): [ 630.365422] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.365693] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] self.force_reraise() [ 630.365693] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.365693] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] raise self.value [ 630.365693] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 630.365693] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] updated_port = self._update_port( [ 630.365693] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.365693] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] _ensure_no_port_binding_failure(port) [ 630.365693] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.365693] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] raise exception.PortBindingFailed(port_id=port['id']) [ 630.365693] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] nova.exception.PortBindingFailed: Binding failed for port fa2b5a1c-551c-44cc-bbf8-8f23c6f3e8ab, please check neutron logs for more information. [ 630.365693] env[61906]: ERROR nova.compute.manager [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] [ 630.366181] env[61906]: DEBUG nova.compute.utils [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Binding failed for port fa2b5a1c-551c-44cc-bbf8-8f23c6f3e8ab, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 630.368479] env[61906]: DEBUG nova.compute.manager [req-287397e7-d96a-4b55-ae8e-aca47b5c2d0e req-45905817-94c2-4bf5-8e83-8acb703ed0fb service nova] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Received event network-vif-deleted-bbd06d6f-2ab5-40f7-ba65-fe54621ee57b {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 630.369086] env[61906]: DEBUG nova.compute.manager [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Build of instance 7ed5b09c-5ba1-479c-96b4-75924df84e75 was re-scheduled: Binding failed for port fa2b5a1c-551c-44cc-bbf8-8f23c6f3e8ab, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 630.369591] env[61906]: DEBUG nova.compute.manager [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 630.369877] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Acquiring lock "refresh_cache-7ed5b09c-5ba1-479c-96b4-75924df84e75" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.370079] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Acquired lock "refresh_cache-7ed5b09c-5ba1-479c-96b4-75924df84e75" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.370330] env[61906]: DEBUG nova.network.neutron [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 630.372367] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.126s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.606135] env[61906]: ERROR nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 85d0c268-50f0-4953-bc6d-34eba905c8fd, please check neutron logs for more information. [ 630.606135] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 630.606135] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.606135] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 630.606135] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 630.606135] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 630.606135] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 630.606135] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 630.606135] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.606135] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 630.606135] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.606135] env[61906]: ERROR nova.compute.manager raise self.value [ 630.606135] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 630.606135] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 630.606135] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.606135] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 630.606576] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.606576] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 630.606576] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 85d0c268-50f0-4953-bc6d-34eba905c8fd, please check neutron logs for more information. [ 630.606576] env[61906]: ERROR nova.compute.manager [ 630.606576] env[61906]: Traceback (most recent call last): [ 630.606576] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 630.606576] env[61906]: listener.cb(fileno) [ 630.606576] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.606576] env[61906]: result = function(*args, **kwargs) [ 630.606576] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 630.606576] env[61906]: return func(*args, **kwargs) [ 630.606576] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 630.606576] env[61906]: raise e [ 630.606576] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.606576] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 630.606576] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 630.606576] env[61906]: created_port_ids = self._update_ports_for_instance( [ 630.606576] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 630.606576] env[61906]: with excutils.save_and_reraise_exception(): [ 630.606576] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.606576] env[61906]: self.force_reraise() [ 630.606576] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.606576] env[61906]: raise self.value [ 630.606576] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 630.606576] env[61906]: updated_port = self._update_port( [ 630.606576] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.606576] env[61906]: _ensure_no_port_binding_failure(port) [ 630.606576] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.606576] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 630.607283] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 85d0c268-50f0-4953-bc6d-34eba905c8fd, please check neutron logs for more information. [ 630.607283] env[61906]: Removing descriptor: 16 [ 630.607283] env[61906]: ERROR nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 85d0c268-50f0-4953-bc6d-34eba905c8fd, please check neutron logs for more information. [ 630.607283] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Traceback (most recent call last): [ 630.607283] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 630.607283] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] yield resources [ 630.607283] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 630.607283] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] self.driver.spawn(context, instance, image_meta, [ 630.607283] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 630.607283] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] self._vmops.spawn(context, instance, image_meta, injected_files, [ 630.607283] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 630.607283] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] vm_ref = self.build_virtual_machine(instance, [ 630.607595] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 630.607595] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] vif_infos = vmwarevif.get_vif_info(self._session, [ 630.607595] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 630.607595] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] for vif in network_info: [ 630.607595] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 630.607595] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] return self._sync_wrapper(fn, *args, **kwargs) [ 630.607595] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 630.607595] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] self.wait() [ 630.607595] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 630.607595] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] self[:] = self._gt.wait() [ 630.607595] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 630.607595] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] return self._exit_event.wait() [ 630.607595] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 630.608107] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] result = hub.switch() [ 630.608107] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 630.608107] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] return self.greenlet.switch() [ 630.608107] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.608107] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] result = function(*args, **kwargs) [ 630.608107] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 630.608107] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] return func(*args, **kwargs) [ 630.608107] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 630.608107] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] raise e [ 630.608107] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.608107] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] nwinfo = self.network_api.allocate_for_instance( [ 630.608107] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 630.608107] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] created_port_ids = self._update_ports_for_instance( [ 630.608490] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 630.608490] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] with excutils.save_and_reraise_exception(): [ 630.608490] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.608490] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] self.force_reraise() [ 630.608490] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.608490] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] raise self.value [ 630.608490] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 630.608490] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] updated_port = self._update_port( [ 630.608490] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.608490] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] _ensure_no_port_binding_failure(port) [ 630.608490] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.608490] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] raise exception.PortBindingFailed(port_id=port['id']) [ 630.608793] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] nova.exception.PortBindingFailed: Binding failed for port 85d0c268-50f0-4953-bc6d-34eba905c8fd, please check neutron logs for more information. [ 630.608793] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] [ 630.610069] env[61906]: INFO nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Terminating instance [ 630.612628] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.624082] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquiring lock "refresh_cache-5b5e9435-ab0c-493a-8ddf-a480960a3818" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.624305] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquired lock "refresh_cache-5b5e9435-ab0c-493a-8ddf-a480960a3818" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.624484] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 630.628258] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Releasing lock "refresh_cache-b4df3af4-f60d-4875-af62-d9c162e4a5c8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 630.628645] env[61906]: DEBUG nova.compute.manager [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 630.628829] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 630.629117] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-41aae048-75a8-4238-a547-3674cf6be856 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.642224] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0007b95c-fc5c-4701-a0bb-356af737fe68 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.670247] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b4df3af4-f60d-4875-af62-d9c162e4a5c8 could not be found. [ 630.670665] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 630.670933] env[61906]: INFO nova.compute.manager [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 630.671273] env[61906]: DEBUG oslo.service.loopingcall [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 630.671591] env[61906]: DEBUG nova.compute.manager [-] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 630.671761] env[61906]: DEBUG nova.network.neutron [-] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 630.687807] env[61906]: DEBUG nova.network.neutron [-] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.910032] env[61906]: DEBUG nova.network.neutron [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.023613] env[61906]: DEBUG nova.network.neutron [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.160589] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.193198] env[61906]: DEBUG nova.network.neutron [-] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.250397] env[61906]: DEBUG oslo_concurrency.lockutils [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Acquiring lock "318a3224-a156-4354-a73b-460168c0eb9c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.251129] env[61906]: DEBUG oslo_concurrency.lockutils [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Lock "318a3224-a156-4354-a73b-460168c0eb9c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.251129] env[61906]: DEBUG oslo_concurrency.lockutils [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Acquiring lock "318a3224-a156-4354-a73b-460168c0eb9c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.251462] env[61906]: DEBUG oslo_concurrency.lockutils [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Lock "318a3224-a156-4354-a73b-460168c0eb9c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.251564] env[61906]: DEBUG oslo_concurrency.lockutils [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Lock "318a3224-a156-4354-a73b-460168c0eb9c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.253584] env[61906]: INFO nova.compute.manager [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Terminating instance [ 631.255218] env[61906]: DEBUG oslo_concurrency.lockutils [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Acquiring lock "refresh_cache-318a3224-a156-4354-a73b-460168c0eb9c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.255430] env[61906]: DEBUG oslo_concurrency.lockutils [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Acquired lock "refresh_cache-318a3224-a156-4354-a73b-460168c0eb9c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.255625] env[61906]: DEBUG nova.network.neutron [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 631.301015] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.466232] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc7de268-8d60-406a-af70-a000ed018383 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.473369] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ca1bfb9-18b9-4c19-b1ca-52cc0a810452 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.508494] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d70ef22d-4468-4a7b-bb39-5b0dc9f2cb92 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.515334] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ef75eae-a810-41c5-8970-5d642efd7121 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.529016] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Releasing lock "refresh_cache-7ed5b09c-5ba1-479c-96b4-75924df84e75" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.529760] env[61906]: DEBUG nova.compute.manager [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 631.529760] env[61906]: DEBUG nova.compute.manager [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 631.529760] env[61906]: DEBUG nova.network.neutron [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 631.531449] env[61906]: DEBUG nova.compute.provider_tree [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 631.697650] env[61906]: INFO nova.compute.manager [-] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Took 1.03 seconds to deallocate network for instance. [ 631.699956] env[61906]: DEBUG nova.compute.claims [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 631.700170] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.701832] env[61906]: DEBUG nova.network.neutron [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.784504] env[61906]: DEBUG nova.network.neutron [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.805234] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Releasing lock "refresh_cache-5b5e9435-ab0c-493a-8ddf-a480960a3818" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.805668] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 631.805842] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 631.806097] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4695f3d9-09a1-4883-b67b-031ce059c111 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.815081] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0121c28f-bc28-43cd-a71d-50e7bde83910 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.838111] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5b5e9435-ab0c-493a-8ddf-a480960a3818 could not be found. [ 631.838111] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 631.838354] env[61906]: INFO nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Took 0.03 seconds to destroy the instance on the hypervisor. [ 631.838519] env[61906]: DEBUG oslo.service.loopingcall [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 631.838734] env[61906]: DEBUG nova.compute.manager [-] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 631.838827] env[61906]: DEBUG nova.network.neutron [-] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 631.855356] env[61906]: DEBUG nova.network.neutron [-] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.885786] env[61906]: DEBUG nova.network.neutron [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.042652] env[61906]: DEBUG nova.scheduler.client.report [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 632.205786] env[61906]: DEBUG nova.network.neutron [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.358540] env[61906]: DEBUG nova.network.neutron [-] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.388688] env[61906]: DEBUG oslo_concurrency.lockutils [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Releasing lock "refresh_cache-318a3224-a156-4354-a73b-460168c0eb9c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.389963] env[61906]: DEBUG nova.compute.manager [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 632.390519] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 632.391882] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1f53015-5665-498a-8d3e-9fee06ff8496 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.402345] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 632.403460] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0cf9925c-99db-457c-8234-4e88cdac3c0b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.412206] env[61906]: DEBUG oslo_vmware.api [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for the task: (returnval){ [ 632.412206] env[61906]: value = "task-1333113" [ 632.412206] env[61906]: _type = "Task" [ 632.412206] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.420012] env[61906]: DEBUG oslo_vmware.api [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333113, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.442213] env[61906]: DEBUG nova.compute.manager [req-40eb133d-df4d-4d46-9f60-bcadc15497f5 req-f0fafe58-25a8-49d2-b6e1-9f90f8ccc9c2 service nova] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Received event network-changed-85d0c268-50f0-4953-bc6d-34eba905c8fd {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 632.442213] env[61906]: DEBUG nova.compute.manager [req-40eb133d-df4d-4d46-9f60-bcadc15497f5 req-f0fafe58-25a8-49d2-b6e1-9f90f8ccc9c2 service nova] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Refreshing instance network info cache due to event network-changed-85d0c268-50f0-4953-bc6d-34eba905c8fd. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 632.442213] env[61906]: DEBUG oslo_concurrency.lockutils [req-40eb133d-df4d-4d46-9f60-bcadc15497f5 req-f0fafe58-25a8-49d2-b6e1-9f90f8ccc9c2 service nova] Acquiring lock "refresh_cache-5b5e9435-ab0c-493a-8ddf-a480960a3818" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.442213] env[61906]: DEBUG oslo_concurrency.lockutils [req-40eb133d-df4d-4d46-9f60-bcadc15497f5 req-f0fafe58-25a8-49d2-b6e1-9f90f8ccc9c2 service nova] Acquired lock "refresh_cache-5b5e9435-ab0c-493a-8ddf-a480960a3818" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.442213] env[61906]: DEBUG nova.network.neutron [req-40eb133d-df4d-4d46-9f60-bcadc15497f5 req-f0fafe58-25a8-49d2-b6e1-9f90f8ccc9c2 service nova] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Refreshing network info cache for port 85d0c268-50f0-4953-bc6d-34eba905c8fd {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 632.549028] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.175s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.549028] env[61906]: ERROR nova.compute.manager [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c43c348f-e3b1-48e5-911e-643fcc5dcef6, please check neutron logs for more information. [ 632.549028] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Traceback (most recent call last): [ 632.549028] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 632.549028] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] self.driver.spawn(context, instance, image_meta, [ 632.549028] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 632.549028] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] self._vmops.spawn(context, instance, image_meta, injected_files, [ 632.549028] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 632.549028] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] vm_ref = self.build_virtual_machine(instance, [ 632.549380] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 632.549380] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] vif_infos = vmwarevif.get_vif_info(self._session, [ 632.549380] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 632.549380] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] for vif in network_info: [ 632.549380] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 632.549380] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] return self._sync_wrapper(fn, *args, **kwargs) [ 632.549380] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 632.549380] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] self.wait() [ 632.549380] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 632.549380] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] self[:] = self._gt.wait() [ 632.549380] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 632.549380] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] return self._exit_event.wait() [ 632.549380] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 632.549699] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] result = hub.switch() [ 632.549699] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 632.549699] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] return self.greenlet.switch() [ 632.549699] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.549699] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] result = function(*args, **kwargs) [ 632.549699] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 632.549699] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] return func(*args, **kwargs) [ 632.549699] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 632.549699] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] raise e [ 632.549699] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.549699] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] nwinfo = self.network_api.allocate_for_instance( [ 632.549699] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.549699] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] created_port_ids = self._update_ports_for_instance( [ 632.550047] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.550047] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] with excutils.save_and_reraise_exception(): [ 632.550047] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.550047] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] self.force_reraise() [ 632.550047] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.550047] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] raise self.value [ 632.550047] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.550047] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] updated_port = self._update_port( [ 632.550047] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.550047] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] _ensure_no_port_binding_failure(port) [ 632.550047] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.550047] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] raise exception.PortBindingFailed(port_id=port['id']) [ 632.550320] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] nova.exception.PortBindingFailed: Binding failed for port c43c348f-e3b1-48e5-911e-643fcc5dcef6, please check neutron logs for more information. [ 632.550320] env[61906]: ERROR nova.compute.manager [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] [ 632.550320] env[61906]: DEBUG nova.compute.utils [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Binding failed for port c43c348f-e3b1-48e5-911e-643fcc5dcef6, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 632.553060] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.698s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.553938] env[61906]: INFO nova.compute.claims [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 632.561131] env[61906]: DEBUG nova.compute.manager [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Build of instance a9249055-f6bf-4693-b0f7-5d8b63510a35 was re-scheduled: Binding failed for port c43c348f-e3b1-48e5-911e-643fcc5dcef6, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 632.561759] env[61906]: DEBUG nova.compute.manager [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 632.566435] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Acquiring lock "refresh_cache-a9249055-f6bf-4693-b0f7-5d8b63510a35" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.566435] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Acquired lock "refresh_cache-a9249055-f6bf-4693-b0f7-5d8b63510a35" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.566435] env[61906]: DEBUG nova.network.neutron [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 632.709186] env[61906]: INFO nova.compute.manager [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] [instance: 7ed5b09c-5ba1-479c-96b4-75924df84e75] Took 1.18 seconds to deallocate network for instance. [ 632.864215] env[61906]: INFO nova.compute.manager [-] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Took 1.03 seconds to deallocate network for instance. [ 632.867061] env[61906]: DEBUG nova.compute.claims [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 632.867061] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.919452] env[61906]: DEBUG oslo_vmware.api [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333113, 'name': PowerOffVM_Task, 'duration_secs': 0.179937} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.919841] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 632.920144] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 632.920476] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-52eca951-3374-4afe-bd69-ff85cb158664 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.944167] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 632.944567] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 632.944892] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Deleting the datastore file [datastore2] 318a3224-a156-4354-a73b-460168c0eb9c {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 632.948280] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a0391872-401b-4a0e-97c6-3ed79ca6e223 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.955186] env[61906]: DEBUG oslo_vmware.api [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for the task: (returnval){ [ 632.955186] env[61906]: value = "task-1333115" [ 632.955186] env[61906]: _type = "Task" [ 632.955186] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.966092] env[61906]: DEBUG oslo_vmware.api [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333115, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.967261] env[61906]: DEBUG nova.network.neutron [req-40eb133d-df4d-4d46-9f60-bcadc15497f5 req-f0fafe58-25a8-49d2-b6e1-9f90f8ccc9c2 service nova] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.061099] env[61906]: DEBUG nova.network.neutron [req-40eb133d-df4d-4d46-9f60-bcadc15497f5 req-f0fafe58-25a8-49d2-b6e1-9f90f8ccc9c2 service nova] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.074022] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 633.074022] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 633.098262] env[61906]: DEBUG nova.network.neutron [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.186788] env[61906]: DEBUG nova.network.neutron [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.467813] env[61906]: DEBUG oslo_vmware.api [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Task: {'id': task-1333115, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.094508} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.470307] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 633.470534] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 633.470746] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 633.471010] env[61906]: INFO nova.compute.manager [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Took 1.08 seconds to destroy the instance on the hypervisor. [ 633.473216] env[61906]: DEBUG oslo.service.loopingcall [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 633.473216] env[61906]: DEBUG nova.compute.manager [-] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 633.473389] env[61906]: DEBUG nova.network.neutron [-] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 633.497500] env[61906]: DEBUG nova.network.neutron [-] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.564215] env[61906]: DEBUG oslo_concurrency.lockutils [req-40eb133d-df4d-4d46-9f60-bcadc15497f5 req-f0fafe58-25a8-49d2-b6e1-9f90f8ccc9c2 service nova] Releasing lock "refresh_cache-5b5e9435-ab0c-493a-8ddf-a480960a3818" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.564414] env[61906]: DEBUG nova.compute.manager [req-40eb133d-df4d-4d46-9f60-bcadc15497f5 req-f0fafe58-25a8-49d2-b6e1-9f90f8ccc9c2 service nova] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Received event network-vif-deleted-85d0c268-50f0-4953-bc6d-34eba905c8fd {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 633.583533] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 633.583533] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Starting heal instance info cache {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 633.583533] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Rebuilding the list of instances to heal {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 633.690917] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Releasing lock "refresh_cache-a9249055-f6bf-4693-b0f7-5d8b63510a35" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.690917] env[61906]: DEBUG nova.compute.manager [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 633.690917] env[61906]: DEBUG nova.compute.manager [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 633.691089] env[61906]: DEBUG nova.network.neutron [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 633.706255] env[61906]: DEBUG nova.network.neutron [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.740399] env[61906]: INFO nova.scheduler.client.report [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Deleted allocations for instance 7ed5b09c-5ba1-479c-96b4-75924df84e75 [ 634.001123] env[61906]: DEBUG nova.network.neutron [-] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.081466] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ff0be9-84aa-48db-9496-0042be5550e7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.090862] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Skipping network cache update for instance because it is being deleted. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 634.090862] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 634.090862] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 634.090862] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 634.090862] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 634.091128] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 634.091128] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 634.101925] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8d71a0d-3421-4825-9119-c6b63774dbf6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.141031] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "refresh_cache-eb4cac2e-85d8-46b6-a4d4-ff12ee70e435" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.141299] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquired lock "refresh_cache-eb4cac2e-85d8-46b6-a4d4-ff12ee70e435" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.142554] env[61906]: DEBUG nova.network.neutron [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Forcefully refreshing network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 634.142554] env[61906]: DEBUG nova.objects.instance [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lazy-loading 'info_cache' on Instance uuid eb4cac2e-85d8-46b6-a4d4-ff12ee70e435 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 634.144387] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5be0efe9-1177-4ee0-ae31-9ab1271b6382 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.155018] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c299526-ef3c-4060-892d-7673c8d7ddc3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.166813] env[61906]: DEBUG nova.compute.provider_tree [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 634.211028] env[61906]: DEBUG nova.network.neutron [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.252334] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b0efee12-136e-4331-91b1-fe73b0dfe01f tempest-ServerDiagnosticsTest-144534735 tempest-ServerDiagnosticsTest-144534735-project-member] Lock "7ed5b09c-5ba1-479c-96b4-75924df84e75" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.679s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.503972] env[61906]: INFO nova.compute.manager [-] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Took 1.03 seconds to deallocate network for instance. [ 634.670375] env[61906]: DEBUG nova.scheduler.client.report [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 634.713769] env[61906]: INFO nova.compute.manager [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] [instance: a9249055-f6bf-4693-b0f7-5d8b63510a35] Took 1.02 seconds to deallocate network for instance. [ 634.758367] env[61906]: DEBUG nova.compute.manager [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 635.010930] env[61906]: DEBUG oslo_concurrency.lockutils [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.173232] env[61906]: DEBUG nova.network.neutron [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 635.177803] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.626s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.178376] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 635.185206] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.381s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.285430] env[61906]: DEBUG oslo_concurrency.lockutils [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.685018] env[61906]: DEBUG nova.compute.utils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 635.685231] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 635.685417] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 635.737297] env[61906]: DEBUG nova.network.neutron [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.739582] env[61906]: DEBUG nova.policy [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d720f18503049d784a1cfe61f242e18', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7a83550c66b84235987277092e630b74', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 635.752300] env[61906]: INFO nova.scheduler.client.report [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Deleted allocations for instance a9249055-f6bf-4693-b0f7-5d8b63510a35 [ 636.129029] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Successfully created port: cd9987a6-8381-4928-a36b-b5d1193bf769 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 636.184417] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a117d90e-0276-45e1-a1c1-289bd009ca3d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.188856] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 636.197539] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28f6335a-f601-40ab-8bf5-1ea61a1f9036 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.228319] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad894655-584f-4ead-9680-05e698db0803 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.235852] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f4c0c0b-cf5c-45e5-816d-ce86eb424d1c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.242765] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Releasing lock "refresh_cache-eb4cac2e-85d8-46b6-a4d4-ff12ee70e435" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.242863] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Updated the network info_cache for instance {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 636.243027] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 636.243591] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 636.854805] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 636.855580] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7322ed1c-4dbb-4eb1-aae5-792d8ea78a7a tempest-ServersTestManualDisk-1982145991 tempest-ServersTestManualDisk-1982145991-project-member] Lock "a9249055-f6bf-4693-b0f7-5d8b63510a35" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.079s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.860411] env[61906]: DEBUG nova.compute.provider_tree [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 636.861672] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 636.865057] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 636.865269] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Acquiring lock "1b289a3d-7958-492b-ab9d-4ba8d3ffdae8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.865466] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Lock "1b289a3d-7958-492b-ab9d-4ba8d3ffdae8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.865734] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 636.865805] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61906) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 636.865889] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 637.065278] env[61906]: DEBUG nova.compute.manager [req-4b425e2f-63b5-4350-b5c4-1bd73aec0b70 req-433f4b9f-d1b9-4e3b-a782-a15afd933688 service nova] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Received event network-changed-cd9987a6-8381-4928-a36b-b5d1193bf769 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 637.065497] env[61906]: DEBUG nova.compute.manager [req-4b425e2f-63b5-4350-b5c4-1bd73aec0b70 req-433f4b9f-d1b9-4e3b-a782-a15afd933688 service nova] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Refreshing instance network info cache due to event network-changed-cd9987a6-8381-4928-a36b-b5d1193bf769. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 637.065709] env[61906]: DEBUG oslo_concurrency.lockutils [req-4b425e2f-63b5-4350-b5c4-1bd73aec0b70 req-433f4b9f-d1b9-4e3b-a782-a15afd933688 service nova] Acquiring lock "refresh_cache-cf331f9d-02ac-45d4-b938-0ed13c63e87b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.065847] env[61906]: DEBUG oslo_concurrency.lockutils [req-4b425e2f-63b5-4350-b5c4-1bd73aec0b70 req-433f4b9f-d1b9-4e3b-a782-a15afd933688 service nova] Acquired lock "refresh_cache-cf331f9d-02ac-45d4-b938-0ed13c63e87b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.066008] env[61906]: DEBUG nova.network.neutron [req-4b425e2f-63b5-4350-b5c4-1bd73aec0b70 req-433f4b9f-d1b9-4e3b-a782-a15afd933688 service nova] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Refreshing network info cache for port cd9987a6-8381-4928-a36b-b5d1193bf769 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 637.363678] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 637.368957] env[61906]: DEBUG nova.compute.manager [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 637.374784] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.375895] env[61906]: DEBUG nova.scheduler.client.report [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 637.410337] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 637.410597] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 637.410760] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 637.411168] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 637.411551] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 637.411896] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 637.411969] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 637.412123] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 637.412320] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 637.412518] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 637.412705] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 637.413779] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b98c0dbd-531c-46ec-9c0c-303e92e18bc2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.417688] env[61906]: ERROR nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cd9987a6-8381-4928-a36b-b5d1193bf769, please check neutron logs for more information. [ 637.417688] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 637.417688] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.417688] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 637.417688] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.417688] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 637.417688] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.417688] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 637.417688] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.417688] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 637.417688] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.417688] env[61906]: ERROR nova.compute.manager raise self.value [ 637.417688] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.417688] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 637.417688] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.417688] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 637.418462] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.418462] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 637.418462] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cd9987a6-8381-4928-a36b-b5d1193bf769, please check neutron logs for more information. [ 637.418462] env[61906]: ERROR nova.compute.manager [ 637.418462] env[61906]: Traceback (most recent call last): [ 637.418462] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 637.418462] env[61906]: listener.cb(fileno) [ 637.418462] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.418462] env[61906]: result = function(*args, **kwargs) [ 637.418462] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 637.418462] env[61906]: return func(*args, **kwargs) [ 637.418462] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 637.418462] env[61906]: raise e [ 637.418462] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.418462] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 637.418462] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.418462] env[61906]: created_port_ids = self._update_ports_for_instance( [ 637.418462] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.418462] env[61906]: with excutils.save_and_reraise_exception(): [ 637.418462] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.418462] env[61906]: self.force_reraise() [ 637.418462] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.418462] env[61906]: raise self.value [ 637.418462] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.418462] env[61906]: updated_port = self._update_port( [ 637.418462] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.418462] env[61906]: _ensure_no_port_binding_failure(port) [ 637.418462] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.418462] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 637.419561] env[61906]: nova.exception.PortBindingFailed: Binding failed for port cd9987a6-8381-4928-a36b-b5d1193bf769, please check neutron logs for more information. [ 637.419561] env[61906]: Removing descriptor: 18 [ 637.423439] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-422d47bb-9eb0-4fb4-88c5-938655b9c97a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.441825] env[61906]: ERROR nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cd9987a6-8381-4928-a36b-b5d1193bf769, please check neutron logs for more information. [ 637.441825] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Traceback (most recent call last): [ 637.441825] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 637.441825] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] yield resources [ 637.441825] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 637.441825] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] self.driver.spawn(context, instance, image_meta, [ 637.441825] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 637.441825] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 637.441825] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 637.441825] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] vm_ref = self.build_virtual_machine(instance, [ 637.441825] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 637.442126] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] vif_infos = vmwarevif.get_vif_info(self._session, [ 637.442126] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 637.442126] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] for vif in network_info: [ 637.442126] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 637.442126] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] return self._sync_wrapper(fn, *args, **kwargs) [ 637.442126] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 637.442126] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] self.wait() [ 637.442126] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 637.442126] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] self[:] = self._gt.wait() [ 637.442126] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 637.442126] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] return self._exit_event.wait() [ 637.442126] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 637.442126] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] current.throw(*self._exc) [ 637.442428] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.442428] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] result = function(*args, **kwargs) [ 637.442428] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 637.442428] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] return func(*args, **kwargs) [ 637.442428] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 637.442428] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] raise e [ 637.442428] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.442428] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] nwinfo = self.network_api.allocate_for_instance( [ 637.442428] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.442428] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] created_port_ids = self._update_ports_for_instance( [ 637.442428] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.442428] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] with excutils.save_and_reraise_exception(): [ 637.442428] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.442736] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] self.force_reraise() [ 637.442736] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.442736] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] raise self.value [ 637.442736] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.442736] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] updated_port = self._update_port( [ 637.442736] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.442736] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] _ensure_no_port_binding_failure(port) [ 637.442736] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.442736] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] raise exception.PortBindingFailed(port_id=port['id']) [ 637.442736] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] nova.exception.PortBindingFailed: Binding failed for port cd9987a6-8381-4928-a36b-b5d1193bf769, please check neutron logs for more information. [ 637.442736] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] [ 637.443778] env[61906]: INFO nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Terminating instance [ 637.446162] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquiring lock "refresh_cache-cf331f9d-02ac-45d4-b938-0ed13c63e87b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.588336] env[61906]: DEBUG nova.network.neutron [req-4b425e2f-63b5-4350-b5c4-1bd73aec0b70 req-433f4b9f-d1b9-4e3b-a782-a15afd933688 service nova] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.725446] env[61906]: DEBUG nova.network.neutron [req-4b425e2f-63b5-4350-b5c4-1bd73aec0b70 req-433f4b9f-d1b9-4e3b-a782-a15afd933688 service nova] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.882679] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.700s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.883246] env[61906]: ERROR nova.compute.manager [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0206d2c8-6ff4-4df4-a6e2-b6a3d7e4c46a, please check neutron logs for more information. [ 637.883246] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Traceback (most recent call last): [ 637.883246] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 637.883246] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] self.driver.spawn(context, instance, image_meta, [ 637.883246] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 637.883246] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 637.883246] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 637.883246] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] vm_ref = self.build_virtual_machine(instance, [ 637.883246] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 637.883246] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] vif_infos = vmwarevif.get_vif_info(self._session, [ 637.883246] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 637.883513] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] for vif in network_info: [ 637.883513] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 637.883513] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] return self._sync_wrapper(fn, *args, **kwargs) [ 637.883513] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 637.883513] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] self.wait() [ 637.883513] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 637.883513] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] self[:] = self._gt.wait() [ 637.883513] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 637.883513] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] return self._exit_event.wait() [ 637.883513] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 637.883513] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] result = hub.switch() [ 637.883513] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 637.883513] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] return self.greenlet.switch() [ 637.883783] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.883783] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] result = function(*args, **kwargs) [ 637.883783] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 637.883783] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] return func(*args, **kwargs) [ 637.883783] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 637.883783] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] raise e [ 637.883783] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.883783] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] nwinfo = self.network_api.allocate_for_instance( [ 637.883783] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.883783] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] created_port_ids = self._update_ports_for_instance( [ 637.883783] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.883783] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] with excutils.save_and_reraise_exception(): [ 637.883783] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.884092] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] self.force_reraise() [ 637.884092] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.884092] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] raise self.value [ 637.884092] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.884092] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] updated_port = self._update_port( [ 637.884092] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.884092] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] _ensure_no_port_binding_failure(port) [ 637.884092] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.884092] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] raise exception.PortBindingFailed(port_id=port['id']) [ 637.884092] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] nova.exception.PortBindingFailed: Binding failed for port 0206d2c8-6ff4-4df4-a6e2-b6a3d7e4c46a, please check neutron logs for more information. [ 637.884092] env[61906]: ERROR nova.compute.manager [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] [ 637.884342] env[61906]: DEBUG nova.compute.utils [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Binding failed for port 0206d2c8-6ff4-4df4-a6e2-b6a3d7e4c46a, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 637.885460] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.066s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.885877] env[61906]: DEBUG nova.objects.instance [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Lazy-loading 'resources' on Instance uuid eb4cac2e-85d8-46b6-a4d4-ff12ee70e435 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 637.887037] env[61906]: DEBUG nova.compute.manager [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Build of instance bff5e099-f6ad-41bd-a55a-6987310c37cf was re-scheduled: Binding failed for port 0206d2c8-6ff4-4df4-a6e2-b6a3d7e4c46a, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 637.887467] env[61906]: DEBUG nova.compute.manager [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 637.888098] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Acquiring lock "refresh_cache-bff5e099-f6ad-41bd-a55a-6987310c37cf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.888262] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Acquired lock "refresh_cache-bff5e099-f6ad-41bd-a55a-6987310c37cf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.888467] env[61906]: DEBUG nova.network.neutron [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 637.897445] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.232302] env[61906]: DEBUG oslo_concurrency.lockutils [req-4b425e2f-63b5-4350-b5c4-1bd73aec0b70 req-433f4b9f-d1b9-4e3b-a782-a15afd933688 service nova] Releasing lock "refresh_cache-cf331f9d-02ac-45d4-b938-0ed13c63e87b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.233023] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquired lock "refresh_cache-cf331f9d-02ac-45d4-b938-0ed13c63e87b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.233221] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 638.426017] env[61906]: DEBUG nova.network.neutron [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.534755] env[61906]: DEBUG nova.network.neutron [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.768433] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.934101] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.963257] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e21a8872-8bed-4e61-9ec0-540f4e0584e0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.970841] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14e28e5a-e9b9-4a7a-baa9-325391d16c67 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.008271] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3840d17f-2fd2-4a89-8727-573a19440151 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.015956] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d912c6f-9e2d-494c-8c9d-70b0640b574a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.030156] env[61906]: DEBUG nova.compute.provider_tree [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.036988] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Releasing lock "refresh_cache-bff5e099-f6ad-41bd-a55a-6987310c37cf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.036988] env[61906]: DEBUG nova.compute.manager [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 639.037142] env[61906]: DEBUG nova.compute.manager [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 639.037297] env[61906]: DEBUG nova.network.neutron [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 639.061737] env[61906]: DEBUG nova.network.neutron [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.149488] env[61906]: DEBUG nova.compute.manager [req-d90fb368-856f-4a40-ba09-f31d27daf3ea req-4878da22-534d-4ab0-91f5-93c383ac8e8b service nova] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Received event network-vif-deleted-cd9987a6-8381-4928-a36b-b5d1193bf769 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 639.436972] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Releasing lock "refresh_cache-cf331f9d-02ac-45d4-b938-0ed13c63e87b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.437467] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 639.438130] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 639.438466] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-031324cf-dbdd-412f-abe3-11b494936205 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.447966] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f929207-8f03-4d43-b406-b996495a5942 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.470638] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cf331f9d-02ac-45d4-b938-0ed13c63e87b could not be found. [ 639.470872] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 639.471069] env[61906]: INFO nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 639.471342] env[61906]: DEBUG oslo.service.loopingcall [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 639.471571] env[61906]: DEBUG nova.compute.manager [-] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 639.471668] env[61906]: DEBUG nova.network.neutron [-] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 639.513333] env[61906]: DEBUG nova.network.neutron [-] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.534286] env[61906]: DEBUG nova.scheduler.client.report [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 639.564087] env[61906]: DEBUG nova.network.neutron [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.016598] env[61906]: DEBUG nova.network.neutron [-] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.043451] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.158s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.046275] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.691s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.068106] env[61906]: INFO nova.compute.manager [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: bff5e099-f6ad-41bd-a55a-6987310c37cf] Took 1.03 seconds to deallocate network for instance. [ 640.076017] env[61906]: INFO nova.scheduler.client.report [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Deleted allocations for instance eb4cac2e-85d8-46b6-a4d4-ff12ee70e435 [ 640.519908] env[61906]: INFO nova.compute.manager [-] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Took 1.05 seconds to deallocate network for instance. [ 640.522737] env[61906]: DEBUG nova.compute.claims [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 640.522898] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.584797] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97638640-d978-408e-9062-e9f80f31bebf tempest-ServerDiagnosticsV248Test-246066395 tempest-ServerDiagnosticsV248Test-246066395-project-member] Lock "eb4cac2e-85d8-46b6-a4d4-ff12ee70e435" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.490s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.051245] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0085c00-a068-45a8-8338-29e8115a3366 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.058998] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e4bbee6-56de-482b-9345-8d3cdcb4ea6a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.091825] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12159fb0-b53b-48a5-9f0a-c19ad65e27fc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.103880] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed47438-2d5a-49af-bac8-366e93cc244b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.119938] env[61906]: DEBUG nova.compute.provider_tree [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.125070] env[61906]: INFO nova.scheduler.client.report [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Deleted allocations for instance bff5e099-f6ad-41bd-a55a-6987310c37cf [ 641.630449] env[61906]: DEBUG nova.scheduler.client.report [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 641.635443] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c54b8d3e-cedb-4382-ab52-eba63fc99833 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Lock "bff5e099-f6ad-41bd-a55a-6987310c37cf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.051s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.838545] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Acquiring lock "830e7d50-5870-4e1f-8f2f-26840ba06d33" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.838545] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Lock "830e7d50-5870-4e1f-8f2f-26840ba06d33" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.139737] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.094s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.140504] env[61906]: ERROR nova.compute.manager [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8b413ae7-89f7-491e-bb59-a9fd1b6547af, please check neutron logs for more information. [ 642.140504] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] Traceback (most recent call last): [ 642.140504] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 642.140504] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] self.driver.spawn(context, instance, image_meta, [ 642.140504] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 642.140504] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] self._vmops.spawn(context, instance, image_meta, injected_files, [ 642.140504] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 642.140504] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] vm_ref = self.build_virtual_machine(instance, [ 642.140504] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 642.140504] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] vif_infos = vmwarevif.get_vif_info(self._session, [ 642.140504] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 642.140793] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] for vif in network_info: [ 642.140793] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 642.140793] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] return self._sync_wrapper(fn, *args, **kwargs) [ 642.140793] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 642.140793] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] self.wait() [ 642.140793] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 642.140793] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] self[:] = self._gt.wait() [ 642.140793] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 642.140793] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] return self._exit_event.wait() [ 642.140793] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 642.140793] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] current.throw(*self._exc) [ 642.140793] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.140793] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] result = function(*args, **kwargs) [ 642.141130] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 642.141130] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] return func(*args, **kwargs) [ 642.141130] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 642.141130] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] raise e [ 642.141130] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.141130] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] nwinfo = self.network_api.allocate_for_instance( [ 642.141130] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 642.141130] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] created_port_ids = self._update_ports_for_instance( [ 642.141130] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 642.141130] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] with excutils.save_and_reraise_exception(): [ 642.141130] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.141130] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] self.force_reraise() [ 642.141130] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.141434] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] raise self.value [ 642.141434] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 642.141434] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] updated_port = self._update_port( [ 642.141434] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.141434] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] _ensure_no_port_binding_failure(port) [ 642.141434] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.141434] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] raise exception.PortBindingFailed(port_id=port['id']) [ 642.141434] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] nova.exception.PortBindingFailed: Binding failed for port 8b413ae7-89f7-491e-bb59-a9fd1b6547af, please check neutron logs for more information. [ 642.141434] env[61906]: ERROR nova.compute.manager [instance: e64791e3-7269-4953-bfe1-fe776744f493] [ 642.141434] env[61906]: DEBUG nova.compute.utils [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Binding failed for port 8b413ae7-89f7-491e-bb59-a9fd1b6547af, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 642.146146] env[61906]: DEBUG nova.compute.manager [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 642.150177] env[61906]: DEBUG nova.compute.manager [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Build of instance e64791e3-7269-4953-bfe1-fe776744f493 was re-scheduled: Binding failed for port 8b413ae7-89f7-491e-bb59-a9fd1b6547af, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 642.150803] env[61906]: DEBUG nova.compute.manager [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 642.151040] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Acquiring lock "refresh_cache-e64791e3-7269-4953-bfe1-fe776744f493" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.151196] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Acquired lock "refresh_cache-e64791e3-7269-4953-bfe1-fe776744f493" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.153086] env[61906]: DEBUG nova.network.neutron [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 642.153821] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 19.939s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.153993] env[61906]: DEBUG nova.objects.instance [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61906) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 642.687481] env[61906]: DEBUG oslo_concurrency.lockutils [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.704321] env[61906]: DEBUG nova.network.neutron [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.853201] env[61906]: DEBUG nova.network.neutron [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.171796] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d1a83858-abd1-4eb0-be16-12f9f7938506 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.172975] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.846s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.360135] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Releasing lock "refresh_cache-e64791e3-7269-4953-bfe1-fe776744f493" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.360135] env[61906]: DEBUG nova.compute.manager [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 643.360135] env[61906]: DEBUG nova.compute.manager [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 643.360135] env[61906]: DEBUG nova.network.neutron [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 643.386121] env[61906]: DEBUG nova.network.neutron [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.883524] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Acquiring lock "5ab321ec-655b-4f68-91aa-9b3db8c692f4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.884134] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Lock "5ab321ec-655b-4f68-91aa-9b3db8c692f4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.890069] env[61906]: DEBUG nova.network.neutron [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.209275] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e1d0f35-92bb-450f-9d13-37ae1e3f2e11 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.218410] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a571f1f-6bca-43a3-9efc-de56ab700541 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.249604] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93cfe37f-9d34-435c-a044-3c753fe5387f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.256215] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bf85f38-e1b6-44d0-8a58-fa06e5b14382 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.270766] env[61906]: DEBUG nova.compute.provider_tree [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 644.392691] env[61906]: INFO nova.compute.manager [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] [instance: e64791e3-7269-4953-bfe1-fe776744f493] Took 1.03 seconds to deallocate network for instance. [ 644.774780] env[61906]: DEBUG nova.scheduler.client.report [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 645.282227] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.107s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.282227] env[61906]: ERROR nova.compute.manager [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4c761c95-c60e-4705-ab93-a0e344175aff, please check neutron logs for more information. [ 645.282227] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Traceback (most recent call last): [ 645.282227] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 645.282227] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] self.driver.spawn(context, instance, image_meta, [ 645.282227] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 645.282227] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] self._vmops.spawn(context, instance, image_meta, injected_files, [ 645.282227] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 645.282227] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] vm_ref = self.build_virtual_machine(instance, [ 645.282703] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 645.282703] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] vif_infos = vmwarevif.get_vif_info(self._session, [ 645.282703] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 645.282703] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] for vif in network_info: [ 645.282703] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 645.282703] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] return self._sync_wrapper(fn, *args, **kwargs) [ 645.282703] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 645.282703] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] self.wait() [ 645.282703] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 645.282703] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] self[:] = self._gt.wait() [ 645.282703] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 645.282703] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] return self._exit_event.wait() [ 645.282703] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 645.283126] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] result = hub.switch() [ 645.283126] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 645.283126] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] return self.greenlet.switch() [ 645.283126] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.283126] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] result = function(*args, **kwargs) [ 645.283126] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 645.283126] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] return func(*args, **kwargs) [ 645.283126] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 645.283126] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] raise e [ 645.283126] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.283126] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] nwinfo = self.network_api.allocate_for_instance( [ 645.283126] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 645.283126] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] created_port_ids = self._update_ports_for_instance( [ 645.283412] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 645.283412] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] with excutils.save_and_reraise_exception(): [ 645.283412] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.283412] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] self.force_reraise() [ 645.283412] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.283412] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] raise self.value [ 645.283412] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 645.283412] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] updated_port = self._update_port( [ 645.283412] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.283412] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] _ensure_no_port_binding_failure(port) [ 645.283412] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.283412] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] raise exception.PortBindingFailed(port_id=port['id']) [ 645.283730] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] nova.exception.PortBindingFailed: Binding failed for port 4c761c95-c60e-4705-ab93-a0e344175aff, please check neutron logs for more information. [ 645.283730] env[61906]: ERROR nova.compute.manager [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] [ 645.283730] env[61906]: DEBUG nova.compute.utils [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Binding failed for port 4c761c95-c60e-4705-ab93-a0e344175aff, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 645.283730] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.612s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.284626] env[61906]: INFO nova.compute.claims [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 645.291143] env[61906]: DEBUG nova.compute.manager [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Build of instance 9fe29bcd-a682-4973-9bb8-069adc204086 was re-scheduled: Binding failed for port 4c761c95-c60e-4705-ab93-a0e344175aff, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 645.291143] env[61906]: DEBUG nova.compute.manager [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 645.291143] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Acquiring lock "refresh_cache-9fe29bcd-a682-4973-9bb8-069adc204086" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.291143] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Acquired lock "refresh_cache-9fe29bcd-a682-4973-9bb8-069adc204086" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.291474] env[61906]: DEBUG nova.network.neutron [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 645.453127] env[61906]: INFO nova.scheduler.client.report [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Deleted allocations for instance e64791e3-7269-4953-bfe1-fe776744f493 [ 645.814619] env[61906]: DEBUG nova.network.neutron [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 645.863873] env[61906]: DEBUG nova.network.neutron [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.966139] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d12049f0-a6b2-4f54-89b4-c813e35645df tempest-AttachInterfacesUnderV243Test-1539371106 tempest-AttachInterfacesUnderV243Test-1539371106-project-member] Lock "e64791e3-7269-4953-bfe1-fe776744f493" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.227s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.370618] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Releasing lock "refresh_cache-9fe29bcd-a682-4973-9bb8-069adc204086" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.370991] env[61906]: DEBUG nova.compute.manager [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 646.371620] env[61906]: DEBUG nova.compute.manager [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 646.371620] env[61906]: DEBUG nova.network.neutron [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 646.392132] env[61906]: DEBUG nova.network.neutron [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.472020] env[61906]: DEBUG nova.compute.manager [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 646.760218] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Acquiring lock "7ad136a2-d850-4b68-af0f-73badf65426c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.760716] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Lock "7ad136a2-d850-4b68-af0f-73badf65426c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.775980] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e28bb51e-a14b-4dc5-8948-034968062a19 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.785218] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef75273e-a95e-4f0e-a7df-0aa7289327ef {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.819271] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2296e5e-0732-4739-994e-b7a4cb0fc109 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.827802] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b2a4c2f-9bfd-4371-a8bd-57bd2efac643 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.844708] env[61906]: DEBUG nova.compute.provider_tree [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 646.895260] env[61906]: DEBUG nova.network.neutron [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.990696] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.351070] env[61906]: DEBUG nova.scheduler.client.report [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 647.396814] env[61906]: INFO nova.compute.manager [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] [instance: 9fe29bcd-a682-4973-9bb8-069adc204086] Took 1.03 seconds to deallocate network for instance. [ 647.855352] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.570s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.855352] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 647.856921] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.107s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.858194] env[61906]: INFO nova.compute.claims [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 648.363402] env[61906]: DEBUG nova.compute.utils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 648.366824] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 648.366989] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 648.445753] env[61906]: INFO nova.scheduler.client.report [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Deleted allocations for instance 9fe29bcd-a682-4973-9bb8-069adc204086 [ 648.452586] env[61906]: DEBUG nova.policy [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d720f18503049d784a1cfe61f242e18', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7a83550c66b84235987277092e630b74', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 648.817769] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Successfully created port: c68c78a3-7249-49ae-94de-57569a36a8d7 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 648.867725] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 648.956630] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7e26078-08e6-4781-a71b-ffc7ca0d795a tempest-TenantUsagesTestJSON-139427047 tempest-TenantUsagesTestJSON-139427047-project-member] Lock "9fe29bcd-a682-4973-9bb8-069adc204086" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.076s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.356100] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1df06a6e-8eb4-47c0-8589-aeaec61f183f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.362330] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7095000-e61b-4975-8089-6b52cc69dc7e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.399590] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b2fc227-2d6c-414a-9f47-4772dd3d1576 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.409695] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-934a7505-9523-4283-aac0-e2441768c9fc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.426798] env[61906]: DEBUG nova.compute.provider_tree [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 649.459883] env[61906]: DEBUG nova.compute.manager [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 649.749493] env[61906]: DEBUG nova.compute.manager [req-042e14d6-9ab1-4459-a32b-3f9f096bc463 req-6aacbbb1-8773-49b9-8a25-ada2bf4a9a1f service nova] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Received event network-changed-c68c78a3-7249-49ae-94de-57569a36a8d7 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 649.749860] env[61906]: DEBUG nova.compute.manager [req-042e14d6-9ab1-4459-a32b-3f9f096bc463 req-6aacbbb1-8773-49b9-8a25-ada2bf4a9a1f service nova] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Refreshing instance network info cache due to event network-changed-c68c78a3-7249-49ae-94de-57569a36a8d7. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 649.751435] env[61906]: DEBUG oslo_concurrency.lockutils [req-042e14d6-9ab1-4459-a32b-3f9f096bc463 req-6aacbbb1-8773-49b9-8a25-ada2bf4a9a1f service nova] Acquiring lock "refresh_cache-5eafc757-e09b-4751-ac50-3e918156832b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 649.751435] env[61906]: DEBUG oslo_concurrency.lockutils [req-042e14d6-9ab1-4459-a32b-3f9f096bc463 req-6aacbbb1-8773-49b9-8a25-ada2bf4a9a1f service nova] Acquired lock "refresh_cache-5eafc757-e09b-4751-ac50-3e918156832b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.751435] env[61906]: DEBUG nova.network.neutron [req-042e14d6-9ab1-4459-a32b-3f9f096bc463 req-6aacbbb1-8773-49b9-8a25-ada2bf4a9a1f service nova] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Refreshing network info cache for port c68c78a3-7249-49ae-94de-57569a36a8d7 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 649.905856] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 649.933216] env[61906]: DEBUG nova.scheduler.client.report [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 649.950192] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 649.950449] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 649.950641] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 649.950855] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 649.951119] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 649.951148] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 649.951358] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 649.951519] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 649.951683] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 649.951843] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 649.952033] env[61906]: DEBUG nova.virt.hardware [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 649.953969] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd69418-87ae-45bc-b775-236faa21fe9e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.962868] env[61906]: ERROR nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c68c78a3-7249-49ae-94de-57569a36a8d7, please check neutron logs for more information. [ 649.962868] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 649.962868] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 649.962868] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 649.962868] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.962868] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 649.962868] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.962868] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 649.962868] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.962868] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 649.962868] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.962868] env[61906]: ERROR nova.compute.manager raise self.value [ 649.962868] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.962868] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 649.962868] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.962868] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 649.963354] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.963354] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 649.963354] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c68c78a3-7249-49ae-94de-57569a36a8d7, please check neutron logs for more information. [ 649.963354] env[61906]: ERROR nova.compute.manager [ 649.963354] env[61906]: Traceback (most recent call last): [ 649.963354] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 649.963354] env[61906]: listener.cb(fileno) [ 649.963354] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 649.963354] env[61906]: result = function(*args, **kwargs) [ 649.963354] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 649.963354] env[61906]: return func(*args, **kwargs) [ 649.963354] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 649.963354] env[61906]: raise e [ 649.963354] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 649.963354] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 649.963354] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.963354] env[61906]: created_port_ids = self._update_ports_for_instance( [ 649.963354] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.963354] env[61906]: with excutils.save_and_reraise_exception(): [ 649.963354] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.963354] env[61906]: self.force_reraise() [ 649.963354] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.963354] env[61906]: raise self.value [ 649.963354] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.963354] env[61906]: updated_port = self._update_port( [ 649.963354] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.963354] env[61906]: _ensure_no_port_binding_failure(port) [ 649.963354] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.963354] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 649.963987] env[61906]: nova.exception.PortBindingFailed: Binding failed for port c68c78a3-7249-49ae-94de-57569a36a8d7, please check neutron logs for more information. [ 649.963987] env[61906]: Removing descriptor: 18 [ 649.964368] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26748c81-2d48-4767-91d7-2abbd8238826 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.985211] env[61906]: ERROR nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c68c78a3-7249-49ae-94de-57569a36a8d7, please check neutron logs for more information. [ 649.985211] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Traceback (most recent call last): [ 649.985211] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 649.985211] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] yield resources [ 649.985211] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 649.985211] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] self.driver.spawn(context, instance, image_meta, [ 649.985211] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 649.985211] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 649.985211] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 649.985211] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] vm_ref = self.build_virtual_machine(instance, [ 649.985211] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 649.985479] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] vif_infos = vmwarevif.get_vif_info(self._session, [ 649.985479] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 649.985479] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] for vif in network_info: [ 649.985479] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 649.985479] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] return self._sync_wrapper(fn, *args, **kwargs) [ 649.985479] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 649.985479] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] self.wait() [ 649.985479] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 649.985479] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] self[:] = self._gt.wait() [ 649.985479] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 649.985479] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] return self._exit_event.wait() [ 649.985479] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 649.985479] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] current.throw(*self._exc) [ 649.985765] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 649.985765] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] result = function(*args, **kwargs) [ 649.985765] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 649.985765] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] return func(*args, **kwargs) [ 649.985765] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 649.985765] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] raise e [ 649.985765] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 649.985765] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] nwinfo = self.network_api.allocate_for_instance( [ 649.985765] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.985765] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] created_port_ids = self._update_ports_for_instance( [ 649.985765] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.985765] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] with excutils.save_and_reraise_exception(): [ 649.985765] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.986068] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] self.force_reraise() [ 649.986068] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.986068] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] raise self.value [ 649.986068] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.986068] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] updated_port = self._update_port( [ 649.986068] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.986068] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] _ensure_no_port_binding_failure(port) [ 649.986068] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.986068] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] raise exception.PortBindingFailed(port_id=port['id']) [ 649.986068] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] nova.exception.PortBindingFailed: Binding failed for port c68c78a3-7249-49ae-94de-57569a36a8d7, please check neutron logs for more information. [ 649.986068] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] [ 649.986068] env[61906]: INFO nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Terminating instance [ 649.988463] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquiring lock "refresh_cache-5eafc757-e09b-4751-ac50-3e918156832b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 649.991807] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.249795] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Acquiring lock "7e452659-0a5e-4740-b7ed-1fd7b1917dac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.249795] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Lock "7e452659-0a5e-4740-b7ed-1fd7b1917dac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.273711] env[61906]: DEBUG nova.network.neutron [req-042e14d6-9ab1-4459-a32b-3f9f096bc463 req-6aacbbb1-8773-49b9-8a25-ada2bf4a9a1f service nova] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.370768] env[61906]: DEBUG nova.network.neutron [req-042e14d6-9ab1-4459-a32b-3f9f096bc463 req-6aacbbb1-8773-49b9-8a25-ada2bf4a9a1f service nova] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.439755] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.583s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.440272] env[61906]: DEBUG nova.compute.manager [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 650.442893] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 19.830s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.443043] env[61906]: DEBUG nova.objects.instance [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61906) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 650.874545] env[61906]: DEBUG oslo_concurrency.lockutils [req-042e14d6-9ab1-4459-a32b-3f9f096bc463 req-6aacbbb1-8773-49b9-8a25-ada2bf4a9a1f service nova] Releasing lock "refresh_cache-5eafc757-e09b-4751-ac50-3e918156832b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.874893] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquired lock "refresh_cache-5eafc757-e09b-4751-ac50-3e918156832b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.874987] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 650.947027] env[61906]: DEBUG nova.compute.utils [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 650.952986] env[61906]: DEBUG nova.compute.manager [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 650.952986] env[61906]: DEBUG nova.network.neutron [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 651.021463] env[61906]: DEBUG nova.policy [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c27dd79c49af4940930afc0c319c81c3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '235fbc5b4af543718a21ccfe8db1141e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 651.349324] env[61906]: DEBUG nova.network.neutron [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Successfully created port: aa5a85af-fe8b-4622-97ee-2db24b015683 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 651.399446] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.450709] env[61906]: DEBUG nova.compute.manager [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 651.458011] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a9a8549-798a-4d4d-891c-f382c74ec0a2 tempest-ServersAdmin275Test-1137234131 tempest-ServersAdmin275Test-1137234131-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.459055] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.759s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.560982] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.788164] env[61906]: DEBUG nova.compute.manager [req-5009edcb-9242-4c7e-ad3b-5bde92f682ae req-836471bc-4621-41e4-a96f-51659225389c service nova] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Received event network-vif-deleted-c68c78a3-7249-49ae-94de-57569a36a8d7 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 652.063375] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Releasing lock "refresh_cache-5eafc757-e09b-4751-ac50-3e918156832b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.064536] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 652.064536] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 652.064536] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bf3cda42-5a18-451d-8ff8-8c36e9ca054f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.076552] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0257eb9c-362c-4414-864c-ee4d941d6486 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.099244] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5eafc757-e09b-4751-ac50-3e918156832b could not be found. [ 652.099485] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 652.099667] env[61906]: INFO nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 652.099934] env[61906]: DEBUG oslo.service.loopingcall [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 652.102547] env[61906]: DEBUG nova.compute.manager [-] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 652.102650] env[61906]: DEBUG nova.network.neutron [-] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 652.125418] env[61906]: DEBUG nova.network.neutron [-] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.426010] env[61906]: ERROR nova.compute.manager [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port aa5a85af-fe8b-4622-97ee-2db24b015683, please check neutron logs for more information. [ 652.426010] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 652.426010] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.426010] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 652.426010] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 652.426010] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 652.426010] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 652.426010] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 652.426010] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.426010] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 652.426010] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.426010] env[61906]: ERROR nova.compute.manager raise self.value [ 652.426010] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 652.426010] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 652.426010] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.426010] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 652.426412] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.426412] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 652.426412] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port aa5a85af-fe8b-4622-97ee-2db24b015683, please check neutron logs for more information. [ 652.426412] env[61906]: ERROR nova.compute.manager [ 652.426412] env[61906]: Traceback (most recent call last): [ 652.426412] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 652.426412] env[61906]: listener.cb(fileno) [ 652.426412] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.426412] env[61906]: result = function(*args, **kwargs) [ 652.426412] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 652.426412] env[61906]: return func(*args, **kwargs) [ 652.426412] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 652.426412] env[61906]: raise e [ 652.426412] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.426412] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 652.426412] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 652.426412] env[61906]: created_port_ids = self._update_ports_for_instance( [ 652.426412] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 652.426412] env[61906]: with excutils.save_and_reraise_exception(): [ 652.426412] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.426412] env[61906]: self.force_reraise() [ 652.426412] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.426412] env[61906]: raise self.value [ 652.426412] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 652.426412] env[61906]: updated_port = self._update_port( [ 652.426412] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.426412] env[61906]: _ensure_no_port_binding_failure(port) [ 652.426412] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.426412] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 652.427166] env[61906]: nova.exception.PortBindingFailed: Binding failed for port aa5a85af-fe8b-4622-97ee-2db24b015683, please check neutron logs for more information. [ 652.427166] env[61906]: Removing descriptor: 18 [ 652.460693] env[61906]: DEBUG nova.compute.manager [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 652.465849] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9ca0882-81c6-42b0-8144-ad07beb9bb67 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.473311] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-968d87d6-567a-4c2b-9059-6e79effaa58c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.506748] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-006cdd47-4d81-4dca-ba8e-471802dae949 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.514400] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc9f413f-d955-45cb-a588-efbbd2bc6c9f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.519900] env[61906]: DEBUG nova.virt.hardware [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 652.520203] env[61906]: DEBUG nova.virt.hardware [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 652.520372] env[61906]: DEBUG nova.virt.hardware [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 652.520558] env[61906]: DEBUG nova.virt.hardware [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 652.520704] env[61906]: DEBUG nova.virt.hardware [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 652.520849] env[61906]: DEBUG nova.virt.hardware [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 652.521084] env[61906]: DEBUG nova.virt.hardware [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 652.521248] env[61906]: DEBUG nova.virt.hardware [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 652.521414] env[61906]: DEBUG nova.virt.hardware [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 652.521576] env[61906]: DEBUG nova.virt.hardware [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 652.521746] env[61906]: DEBUG nova.virt.hardware [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 652.522463] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d50c60ce-b42c-4ddd-a595-af86e35ba0f0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.535313] env[61906]: DEBUG nova.compute.provider_tree [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 652.539402] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d01871-ee42-45d7-bc64-fb8081808dd1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.552956] env[61906]: ERROR nova.compute.manager [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port aa5a85af-fe8b-4622-97ee-2db24b015683, please check neutron logs for more information. [ 652.552956] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Traceback (most recent call last): [ 652.552956] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 652.552956] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] yield resources [ 652.552956] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 652.552956] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] self.driver.spawn(context, instance, image_meta, [ 652.552956] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 652.552956] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 652.552956] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 652.552956] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] vm_ref = self.build_virtual_machine(instance, [ 652.552956] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 652.553389] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] vif_infos = vmwarevif.get_vif_info(self._session, [ 652.553389] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 652.553389] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] for vif in network_info: [ 652.553389] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 652.553389] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] return self._sync_wrapper(fn, *args, **kwargs) [ 652.553389] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 652.553389] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] self.wait() [ 652.553389] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 652.553389] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] self[:] = self._gt.wait() [ 652.553389] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 652.553389] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] return self._exit_event.wait() [ 652.553389] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 652.553389] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] current.throw(*self._exc) [ 652.553801] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.553801] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] result = function(*args, **kwargs) [ 652.553801] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 652.553801] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] return func(*args, **kwargs) [ 652.553801] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 652.553801] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] raise e [ 652.553801] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.553801] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] nwinfo = self.network_api.allocate_for_instance( [ 652.553801] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 652.553801] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] created_port_ids = self._update_ports_for_instance( [ 652.553801] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 652.553801] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] with excutils.save_and_reraise_exception(): [ 652.553801] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.555114] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] self.force_reraise() [ 652.555114] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.555114] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] raise self.value [ 652.555114] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 652.555114] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] updated_port = self._update_port( [ 652.555114] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.555114] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] _ensure_no_port_binding_failure(port) [ 652.555114] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.555114] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] raise exception.PortBindingFailed(port_id=port['id']) [ 652.555114] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] nova.exception.PortBindingFailed: Binding failed for port aa5a85af-fe8b-4622-97ee-2db24b015683, please check neutron logs for more information. [ 652.555114] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] [ 652.555114] env[61906]: INFO nova.compute.manager [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Terminating instance [ 652.555387] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Acquiring lock "refresh_cache-3595e6c6-10f3-46ba-85c4-90e25de78c1b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.555387] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Acquired lock "refresh_cache-3595e6c6-10f3-46ba-85c4-90e25de78c1b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.555443] env[61906]: DEBUG nova.network.neutron [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 652.631124] env[61906]: DEBUG nova.network.neutron [-] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.040127] env[61906]: DEBUG nova.scheduler.client.report [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 653.082055] env[61906]: DEBUG nova.network.neutron [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.132488] env[61906]: DEBUG nova.network.neutron [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.137062] env[61906]: INFO nova.compute.manager [-] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Took 1.03 seconds to deallocate network for instance. [ 653.142832] env[61906]: DEBUG nova.compute.claims [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 653.142832] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.545304] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.086s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.546405] env[61906]: ERROR nova.compute.manager [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bbd06d6f-2ab5-40f7-ba65-fe54621ee57b, please check neutron logs for more information. [ 653.546405] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Traceback (most recent call last): [ 653.546405] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 653.546405] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] self.driver.spawn(context, instance, image_meta, [ 653.546405] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 653.546405] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 653.546405] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 653.546405] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] vm_ref = self.build_virtual_machine(instance, [ 653.546405] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 653.546405] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] vif_infos = vmwarevif.get_vif_info(self._session, [ 653.546405] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 653.546739] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] for vif in network_info: [ 653.546739] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 653.546739] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] return self._sync_wrapper(fn, *args, **kwargs) [ 653.546739] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 653.546739] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] self.wait() [ 653.546739] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 653.546739] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] self[:] = self._gt.wait() [ 653.546739] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 653.546739] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] return self._exit_event.wait() [ 653.546739] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 653.546739] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] result = hub.switch() [ 653.546739] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 653.546739] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] return self.greenlet.switch() [ 653.547142] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.547142] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] result = function(*args, **kwargs) [ 653.547142] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 653.547142] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] return func(*args, **kwargs) [ 653.547142] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 653.547142] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] raise e [ 653.547142] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.547142] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] nwinfo = self.network_api.allocate_for_instance( [ 653.547142] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.547142] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] created_port_ids = self._update_ports_for_instance( [ 653.547142] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.547142] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] with excutils.save_and_reraise_exception(): [ 653.547142] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.547508] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] self.force_reraise() [ 653.547508] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.547508] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] raise self.value [ 653.547508] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.547508] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] updated_port = self._update_port( [ 653.547508] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.547508] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] _ensure_no_port_binding_failure(port) [ 653.547508] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.547508] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] raise exception.PortBindingFailed(port_id=port['id']) [ 653.547508] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] nova.exception.PortBindingFailed: Binding failed for port bbd06d6f-2ab5-40f7-ba65-fe54621ee57b, please check neutron logs for more information. [ 653.547508] env[61906]: ERROR nova.compute.manager [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] [ 653.547883] env[61906]: DEBUG nova.compute.utils [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Binding failed for port bbd06d6f-2ab5-40f7-ba65-fe54621ee57b, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 653.548365] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.681s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.551625] env[61906]: DEBUG nova.compute.manager [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Build of instance b4df3af4-f60d-4875-af62-d9c162e4a5c8 was re-scheduled: Binding failed for port bbd06d6f-2ab5-40f7-ba65-fe54621ee57b, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 653.551754] env[61906]: DEBUG nova.compute.manager [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 653.551934] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Acquiring lock "refresh_cache-b4df3af4-f60d-4875-af62-d9c162e4a5c8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.554425] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Acquired lock "refresh_cache-b4df3af4-f60d-4875-af62-d9c162e4a5c8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.554425] env[61906]: DEBUG nova.network.neutron [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 653.642753] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Releasing lock "refresh_cache-3595e6c6-10f3-46ba-85c4-90e25de78c1b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.643219] env[61906]: DEBUG nova.compute.manager [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 653.643419] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 653.643725] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bf0cb8a0-ff47-448f-8542-0ff9712de6cb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.652396] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7646a68-9538-4580-9a19-995f2ed68c61 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.675034] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3595e6c6-10f3-46ba-85c4-90e25de78c1b could not be found. [ 653.675442] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 653.675757] env[61906]: INFO nova.compute.manager [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 653.676138] env[61906]: DEBUG oslo.service.loopingcall [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 653.676478] env[61906]: DEBUG nova.compute.manager [-] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 653.676699] env[61906]: DEBUG nova.network.neutron [-] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 653.692808] env[61906]: DEBUG nova.network.neutron [-] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.816676] env[61906]: DEBUG nova.compute.manager [req-2b1b78a9-2ac1-4137-9087-3ac2dfa018af req-f14788d8-034d-4beb-9ba5-e57235b5c308 service nova] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Received event network-changed-aa5a85af-fe8b-4622-97ee-2db24b015683 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 653.816883] env[61906]: DEBUG nova.compute.manager [req-2b1b78a9-2ac1-4137-9087-3ac2dfa018af req-f14788d8-034d-4beb-9ba5-e57235b5c308 service nova] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Refreshing instance network info cache due to event network-changed-aa5a85af-fe8b-4622-97ee-2db24b015683. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 653.817114] env[61906]: DEBUG oslo_concurrency.lockutils [req-2b1b78a9-2ac1-4137-9087-3ac2dfa018af req-f14788d8-034d-4beb-9ba5-e57235b5c308 service nova] Acquiring lock "refresh_cache-3595e6c6-10f3-46ba-85c4-90e25de78c1b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.817729] env[61906]: DEBUG oslo_concurrency.lockutils [req-2b1b78a9-2ac1-4137-9087-3ac2dfa018af req-f14788d8-034d-4beb-9ba5-e57235b5c308 service nova] Acquired lock "refresh_cache-3595e6c6-10f3-46ba-85c4-90e25de78c1b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.817939] env[61906]: DEBUG nova.network.neutron [req-2b1b78a9-2ac1-4137-9087-3ac2dfa018af req-f14788d8-034d-4beb-9ba5-e57235b5c308 service nova] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Refreshing network info cache for port aa5a85af-fe8b-4622-97ee-2db24b015683 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 653.845860] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Acquiring lock "b935c7ba-7716-443b-b5da-9ee03cef793a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.845860] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Lock "b935c7ba-7716-443b-b5da-9ee03cef793a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.073095] env[61906]: DEBUG nova.network.neutron [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 654.165287] env[61906]: DEBUG nova.network.neutron [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.196249] env[61906]: DEBUG nova.network.neutron [-] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.337848] env[61906]: DEBUG nova.network.neutron [req-2b1b78a9-2ac1-4137-9087-3ac2dfa018af req-f14788d8-034d-4beb-9ba5-e57235b5c308 service nova] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 654.412158] env[61906]: DEBUG nova.network.neutron [req-2b1b78a9-2ac1-4137-9087-3ac2dfa018af req-f14788d8-034d-4beb-9ba5-e57235b5c308 service nova] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.550750] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9cab70d-aab8-4bf9-881e-84f45231dcba {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.558508] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78cb84cf-1494-4138-9b78-d3232c1cf23e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.589178] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62427815-323b-4e48-8d4f-8de82e6e98d7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.599029] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e04f3cf3-9070-4bb7-ae6d-8af128014424 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.610218] env[61906]: DEBUG nova.compute.provider_tree [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 654.667941] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Releasing lock "refresh_cache-b4df3af4-f60d-4875-af62-d9c162e4a5c8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.668299] env[61906]: DEBUG nova.compute.manager [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 654.668490] env[61906]: DEBUG nova.compute.manager [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 654.668725] env[61906]: DEBUG nova.network.neutron [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 654.684045] env[61906]: DEBUG nova.network.neutron [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 654.699945] env[61906]: INFO nova.compute.manager [-] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Took 1.02 seconds to deallocate network for instance. [ 654.702098] env[61906]: DEBUG nova.compute.claims [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 654.702277] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.914688] env[61906]: DEBUG oslo_concurrency.lockutils [req-2b1b78a9-2ac1-4137-9087-3ac2dfa018af req-f14788d8-034d-4beb-9ba5-e57235b5c308 service nova] Releasing lock "refresh_cache-3595e6c6-10f3-46ba-85c4-90e25de78c1b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.915067] env[61906]: DEBUG nova.compute.manager [req-2b1b78a9-2ac1-4137-9087-3ac2dfa018af req-f14788d8-034d-4beb-9ba5-e57235b5c308 service nova] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Received event network-vif-deleted-aa5a85af-fe8b-4622-97ee-2db24b015683 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 655.113081] env[61906]: DEBUG nova.scheduler.client.report [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 655.186301] env[61906]: DEBUG nova.network.neutron [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.618567] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.070s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.619229] env[61906]: ERROR nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 85d0c268-50f0-4953-bc6d-34eba905c8fd, please check neutron logs for more information. [ 655.619229] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Traceback (most recent call last): [ 655.619229] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 655.619229] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] self.driver.spawn(context, instance, image_meta, [ 655.619229] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 655.619229] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] self._vmops.spawn(context, instance, image_meta, injected_files, [ 655.619229] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 655.619229] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] vm_ref = self.build_virtual_machine(instance, [ 655.619229] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 655.619229] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] vif_infos = vmwarevif.get_vif_info(self._session, [ 655.619229] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 655.619573] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] for vif in network_info: [ 655.619573] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 655.619573] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] return self._sync_wrapper(fn, *args, **kwargs) [ 655.619573] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 655.619573] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] self.wait() [ 655.619573] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 655.619573] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] self[:] = self._gt.wait() [ 655.619573] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 655.619573] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] return self._exit_event.wait() [ 655.619573] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 655.619573] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] result = hub.switch() [ 655.619573] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 655.619573] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] return self.greenlet.switch() [ 655.619875] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 655.619875] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] result = function(*args, **kwargs) [ 655.619875] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 655.619875] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] return func(*args, **kwargs) [ 655.619875] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 655.619875] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] raise e [ 655.619875] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 655.619875] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] nwinfo = self.network_api.allocate_for_instance( [ 655.619875] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 655.619875] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] created_port_ids = self._update_ports_for_instance( [ 655.619875] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 655.619875] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] with excutils.save_and_reraise_exception(): [ 655.619875] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.620217] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] self.force_reraise() [ 655.620217] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.620217] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] raise self.value [ 655.620217] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 655.620217] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] updated_port = self._update_port( [ 655.620217] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.620217] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] _ensure_no_port_binding_failure(port) [ 655.620217] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.620217] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] raise exception.PortBindingFailed(port_id=port['id']) [ 655.620217] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] nova.exception.PortBindingFailed: Binding failed for port 85d0c268-50f0-4953-bc6d-34eba905c8fd, please check neutron logs for more information. [ 655.620217] env[61906]: ERROR nova.compute.manager [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] [ 655.620494] env[61906]: DEBUG nova.compute.utils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Binding failed for port 85d0c268-50f0-4953-bc6d-34eba905c8fd, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 655.621394] env[61906]: DEBUG oslo_concurrency.lockutils [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.610s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.621506] env[61906]: DEBUG nova.objects.instance [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Lazy-loading 'resources' on Instance uuid 318a3224-a156-4354-a73b-460168c0eb9c {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 655.622775] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Build of instance 5b5e9435-ab0c-493a-8ddf-a480960a3818 was re-scheduled: Binding failed for port 85d0c268-50f0-4953-bc6d-34eba905c8fd, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 655.623207] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 655.623430] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquiring lock "refresh_cache-5b5e9435-ab0c-493a-8ddf-a480960a3818" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.623574] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquired lock "refresh_cache-5b5e9435-ab0c-493a-8ddf-a480960a3818" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.623732] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 655.689860] env[61906]: INFO nova.compute.manager [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] [instance: b4df3af4-f60d-4875-af62-d9c162e4a5c8] Took 1.02 seconds to deallocate network for instance. [ 656.141023] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.223674] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.499754] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba91cb1-8c3c-42be-bde5-207a01c58645 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.507210] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82aa6c9d-23c0-4ebc-960d-2d404dfde0b3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.536514] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4b50015-477e-4dc1-84c3-bdc03cc002e4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.543779] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ada65b5-0e77-4a48-bab7-beef8116cd4e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.556505] env[61906]: DEBUG nova.compute.provider_tree [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.716083] env[61906]: INFO nova.scheduler.client.report [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Deleted allocations for instance b4df3af4-f60d-4875-af62-d9c162e4a5c8 [ 656.730479] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Releasing lock "refresh_cache-5b5e9435-ab0c-493a-8ddf-a480960a3818" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.730479] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 656.730617] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 656.731526] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 656.749382] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.059565] env[61906]: DEBUG nova.scheduler.client.report [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 657.224786] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e9b132d8-d805-4c0e-8ae1-a8d024ab5936 tempest-AttachInterfacesV270Test-1851804368 tempest-AttachInterfacesV270Test-1851804368-project-member] Lock "b4df3af4-f60d-4875-af62-d9c162e4a5c8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.360s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.251354] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.565106] env[61906]: DEBUG oslo_concurrency.lockutils [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.944s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.567668] env[61906]: DEBUG oslo_concurrency.lockutils [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.282s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.569832] env[61906]: INFO nova.compute.claims [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 657.594376] env[61906]: INFO nova.scheduler.client.report [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Deleted allocations for instance 318a3224-a156-4354-a73b-460168c0eb9c [ 657.727549] env[61906]: DEBUG nova.compute.manager [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 657.755829] env[61906]: INFO nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5b5e9435-ab0c-493a-8ddf-a480960a3818] Took 1.02 seconds to deallocate network for instance. [ 658.107085] env[61906]: DEBUG oslo_concurrency.lockutils [None req-43131522-ffff-4160-8c29-595927cc7b24 tempest-ServersAdmin275Test-856098162 tempest-ServersAdmin275Test-856098162-project-member] Lock "318a3224-a156-4354-a73b-460168c0eb9c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.856s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.253626] env[61906]: DEBUG oslo_concurrency.lockutils [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.800202] env[61906]: INFO nova.scheduler.client.report [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Deleted allocations for instance 5b5e9435-ab0c-493a-8ddf-a480960a3818 [ 659.142182] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bcbc46f-168c-4050-ab00-f8f0a58cf47f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.150239] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60325456-d02d-4208-aba9-9f4e30ddc3f5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.184672] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54c9ee20-2eb4-4878-9319-aedeafa75232 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.196312] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d749fd13-0af5-45ef-8f4b-e3ab3d551d74 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.212281] env[61906]: DEBUG nova.compute.provider_tree [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 659.310968] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "5b5e9435-ab0c-493a-8ddf-a480960a3818" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.088s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.717849] env[61906]: DEBUG nova.scheduler.client.report [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 659.813495] env[61906]: DEBUG nova.compute.manager [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 660.222735] env[61906]: DEBUG oslo_concurrency.lockutils [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.655s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.224680] env[61906]: DEBUG nova.compute.manager [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 660.226418] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 22.852s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.226663] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.226745] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61906) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 660.227040] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.330s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.228498] env[61906]: INFO nova.compute.claims [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 660.231752] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c01f67da-8179-48c5-995a-37b48916f80e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.240362] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-157b50c1-a98d-49cf-a9d4-2da54ee32fe1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.261875] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c88a06d-bc5c-4495-a2e3-9c37e6f1494a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.269030] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07ea8f32-4f67-4c07-8c65-2d99dabf161a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.303270] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181205MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=61906) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 660.303437] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.342893] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.422107] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fcc0788-eb9c-4b86-aa29-521cdd8b524d tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquiring lock "5eafc757-e09b-4751-ac50-3e918156832b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.532487] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bce7d229-d0b5-430a-b2e3-2625c02a7950 tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquiring lock "cf331f9d-02ac-45d4-b938-0ed13c63e87b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.733809] env[61906]: DEBUG nova.compute.utils [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 660.735288] env[61906]: DEBUG nova.compute.manager [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 660.735468] env[61906]: DEBUG nova.network.neutron [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 660.825572] env[61906]: DEBUG nova.policy [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '234a01fb35844605a570abf44b290a04', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a9bf98fbf621473c816aabef49e0f547', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 661.246252] env[61906]: DEBUG nova.compute.manager [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 661.468982] env[61906]: DEBUG nova.network.neutron [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Successfully created port: 9691890e-5d7f-4f1d-910a-a3d195288529 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 661.862993] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-658d2308-346d-4727-a092-0068c840b248 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.870632] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e40b769a-057c-4e65-9234-a0452a9190a4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.910133] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-252b06e9-2a65-43a5-bbf2-d0b15caa66da {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.918128] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44ff2839-cd44-462f-ba78-96bc29cfedaf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.931777] env[61906]: DEBUG nova.compute.provider_tree [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 662.262245] env[61906]: DEBUG nova.compute.manager [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 662.298375] env[61906]: DEBUG nova.virt.hardware [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 662.298523] env[61906]: DEBUG nova.virt.hardware [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 662.298654] env[61906]: DEBUG nova.virt.hardware [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 662.298836] env[61906]: DEBUG nova.virt.hardware [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 662.298980] env[61906]: DEBUG nova.virt.hardware [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 662.299150] env[61906]: DEBUG nova.virt.hardware [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 662.299437] env[61906]: DEBUG nova.virt.hardware [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 662.299512] env[61906]: DEBUG nova.virt.hardware [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 662.300435] env[61906]: DEBUG nova.virt.hardware [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 662.300435] env[61906]: DEBUG nova.virt.hardware [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 662.300435] env[61906]: DEBUG nova.virt.hardware [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 662.301260] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e809f8c3-e14b-44d4-9dc4-4e26f7124dcd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.309506] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d1d8db-d2c5-4a86-a49a-e412368ab4f1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.437146] env[61906]: DEBUG nova.scheduler.client.report [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 662.573284] env[61906]: ERROR nova.compute.manager [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9691890e-5d7f-4f1d-910a-a3d195288529, please check neutron logs for more information. [ 662.573284] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 662.573284] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 662.573284] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 662.573284] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.573284] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 662.573284] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.573284] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 662.573284] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.573284] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 662.573284] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.573284] env[61906]: ERROR nova.compute.manager raise self.value [ 662.573284] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.573284] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 662.573284] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.573284] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 662.573943] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.573943] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 662.573943] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9691890e-5d7f-4f1d-910a-a3d195288529, please check neutron logs for more information. [ 662.573943] env[61906]: ERROR nova.compute.manager [ 662.574508] env[61906]: Traceback (most recent call last): [ 662.574508] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 662.574508] env[61906]: listener.cb(fileno) [ 662.574508] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 662.574508] env[61906]: result = function(*args, **kwargs) [ 662.574508] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 662.574508] env[61906]: return func(*args, **kwargs) [ 662.574508] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 662.574508] env[61906]: raise e [ 662.574508] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 662.574508] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 662.574508] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.574508] env[61906]: created_port_ids = self._update_ports_for_instance( [ 662.574508] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.574508] env[61906]: with excutils.save_and_reraise_exception(): [ 662.574508] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.574508] env[61906]: self.force_reraise() [ 662.574508] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.574508] env[61906]: raise self.value [ 662.574508] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.574508] env[61906]: updated_port = self._update_port( [ 662.574508] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.574508] env[61906]: _ensure_no_port_binding_failure(port) [ 662.574508] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.574508] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 662.574508] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 9691890e-5d7f-4f1d-910a-a3d195288529, please check neutron logs for more information. [ 662.574508] env[61906]: Removing descriptor: 18 [ 662.575299] env[61906]: ERROR nova.compute.manager [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9691890e-5d7f-4f1d-910a-a3d195288529, please check neutron logs for more information. [ 662.575299] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Traceback (most recent call last): [ 662.575299] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 662.575299] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] yield resources [ 662.575299] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 662.575299] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] self.driver.spawn(context, instance, image_meta, [ 662.575299] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 662.575299] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 662.575299] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 662.575299] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] vm_ref = self.build_virtual_machine(instance, [ 662.575299] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 662.575608] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] vif_infos = vmwarevif.get_vif_info(self._session, [ 662.575608] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 662.575608] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] for vif in network_info: [ 662.575608] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 662.575608] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] return self._sync_wrapper(fn, *args, **kwargs) [ 662.575608] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 662.575608] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] self.wait() [ 662.575608] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 662.575608] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] self[:] = self._gt.wait() [ 662.575608] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 662.575608] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] return self._exit_event.wait() [ 662.575608] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 662.575608] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] result = hub.switch() [ 662.575978] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 662.575978] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] return self.greenlet.switch() [ 662.575978] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 662.575978] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] result = function(*args, **kwargs) [ 662.575978] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 662.575978] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] return func(*args, **kwargs) [ 662.575978] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 662.575978] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] raise e [ 662.575978] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 662.575978] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] nwinfo = self.network_api.allocate_for_instance( [ 662.575978] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.575978] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] created_port_ids = self._update_ports_for_instance( [ 662.575978] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.576342] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] with excutils.save_and_reraise_exception(): [ 662.576342] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.576342] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] self.force_reraise() [ 662.576342] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.576342] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] raise self.value [ 662.576342] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.576342] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] updated_port = self._update_port( [ 662.576342] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.576342] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] _ensure_no_port_binding_failure(port) [ 662.576342] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.576342] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] raise exception.PortBindingFailed(port_id=port['id']) [ 662.576342] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] nova.exception.PortBindingFailed: Binding failed for port 9691890e-5d7f-4f1d-910a-a3d195288529, please check neutron logs for more information. [ 662.576342] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] [ 662.576840] env[61906]: INFO nova.compute.manager [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Terminating instance [ 662.577907] env[61906]: DEBUG oslo_concurrency.lockutils [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Acquiring lock "refresh_cache-32e2f482-9ceb-44bc-8933-115876374b1c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.578089] env[61906]: DEBUG oslo_concurrency.lockutils [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Acquired lock "refresh_cache-32e2f482-9ceb-44bc-8933-115876374b1c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.578261] env[61906]: DEBUG nova.network.neutron [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 662.592174] env[61906]: DEBUG nova.compute.manager [req-1af50821-96fd-4687-8564-cfa2905a8019 req-afd498c9-5ca1-4a61-b5f0-d5f3f34065ce service nova] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Received event network-changed-9691890e-5d7f-4f1d-910a-a3d195288529 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 662.592374] env[61906]: DEBUG nova.compute.manager [req-1af50821-96fd-4687-8564-cfa2905a8019 req-afd498c9-5ca1-4a61-b5f0-d5f3f34065ce service nova] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Refreshing instance network info cache due to event network-changed-9691890e-5d7f-4f1d-910a-a3d195288529. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 662.592585] env[61906]: DEBUG oslo_concurrency.lockutils [req-1af50821-96fd-4687-8564-cfa2905a8019 req-afd498c9-5ca1-4a61-b5f0-d5f3f34065ce service nova] Acquiring lock "refresh_cache-32e2f482-9ceb-44bc-8933-115876374b1c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.948630] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.721s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.949847] env[61906]: DEBUG nova.compute.manager [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 662.956180] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.433s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.108288] env[61906]: DEBUG nova.network.neutron [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.283438] env[61906]: DEBUG nova.network.neutron [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.466823] env[61906]: DEBUG nova.compute.utils [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 663.468855] env[61906]: DEBUG nova.compute.manager [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 663.469041] env[61906]: DEBUG nova.network.neutron [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 663.512630] env[61906]: DEBUG nova.policy [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0eec8c5a3d344fd78abf0b9d1353ef8a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f106f4647bae40929c743a6b4be3eafe', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 663.786151] env[61906]: DEBUG oslo_concurrency.lockutils [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Releasing lock "refresh_cache-32e2f482-9ceb-44bc-8933-115876374b1c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.786663] env[61906]: DEBUG nova.compute.manager [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 663.786785] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 663.788407] env[61906]: DEBUG oslo_concurrency.lockutils [req-1af50821-96fd-4687-8564-cfa2905a8019 req-afd498c9-5ca1-4a61-b5f0-d5f3f34065ce service nova] Acquired lock "refresh_cache-32e2f482-9ceb-44bc-8933-115876374b1c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.788591] env[61906]: DEBUG nova.network.neutron [req-1af50821-96fd-4687-8564-cfa2905a8019 req-afd498c9-5ca1-4a61-b5f0-d5f3f34065ce service nova] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Refreshing network info cache for port 9691890e-5d7f-4f1d-910a-a3d195288529 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 663.789633] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-81be1010-5f63-427b-9131-e50ad1eea6b9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.805079] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84e401d8-f187-4f7d-a437-4f36a31c374b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.830396] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 32e2f482-9ceb-44bc-8933-115876374b1c could not be found. [ 663.833479] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 663.833479] env[61906]: INFO nova.compute.manager [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 663.833479] env[61906]: DEBUG oslo.service.loopingcall [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 663.833479] env[61906]: DEBUG nova.compute.manager [-] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 663.833479] env[61906]: DEBUG nova.network.neutron [-] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 663.845266] env[61906]: DEBUG nova.network.neutron [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Successfully created port: 75dff989-71c3-4b84-811d-eb77b29ac72d {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 663.851219] env[61906]: DEBUG nova.network.neutron [-] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.956373] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a00ce21d-293d-446d-a391-01f865a10e18 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.965223] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c66d39c4-eb6c-4d55-9e63-89b63611add3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.020848] env[61906]: DEBUG nova.compute.manager [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 664.027551] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e4efda-7a87-4413-9697-fa60b67f6f9e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.041256] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dc6e053-155d-4f79-b8cc-f0130fb872ae {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.059194] env[61906]: DEBUG nova.compute.provider_tree [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 664.315696] env[61906]: DEBUG nova.network.neutron [req-1af50821-96fd-4687-8564-cfa2905a8019 req-afd498c9-5ca1-4a61-b5f0-d5f3f34065ce service nova] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.357132] env[61906]: DEBUG nova.network.neutron [-] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.433425] env[61906]: DEBUG nova.network.neutron [req-1af50821-96fd-4687-8564-cfa2905a8019 req-afd498c9-5ca1-4a61-b5f0-d5f3f34065ce service nova] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.514150] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Acquiring lock "3923d04e-5b50-4cfc-b3da-4b4eada4c8e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.514150] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Lock "3923d04e-5b50-4cfc-b3da-4b4eada4c8e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.565486] env[61906]: DEBUG nova.scheduler.client.report [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 664.627135] env[61906]: DEBUG nova.compute.manager [req-91c666df-ff9c-40c1-9c1a-5c247d41316d req-9d9d651d-5566-4227-b660-3747a3a3ccab service nova] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Received event network-vif-deleted-9691890e-5d7f-4f1d-910a-a3d195288529 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 664.860310] env[61906]: INFO nova.compute.manager [-] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Took 1.03 seconds to deallocate network for instance. [ 664.862727] env[61906]: DEBUG nova.compute.claims [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 664.862909] env[61906]: DEBUG oslo_concurrency.lockutils [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.934171] env[61906]: DEBUG oslo_concurrency.lockutils [req-1af50821-96fd-4687-8564-cfa2905a8019 req-afd498c9-5ca1-4a61-b5f0-d5f3f34065ce service nova] Releasing lock "refresh_cache-32e2f482-9ceb-44bc-8933-115876374b1c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.986169] env[61906]: ERROR nova.compute.manager [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 75dff989-71c3-4b84-811d-eb77b29ac72d, please check neutron logs for more information. [ 664.986169] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 664.986169] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.986169] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 664.986169] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 664.986169] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 664.986169] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 664.986169] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 664.986169] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.986169] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 664.986169] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.986169] env[61906]: ERROR nova.compute.manager raise self.value [ 664.986169] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 664.986169] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 664.986169] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.986169] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 664.986597] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.986597] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 664.986597] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 75dff989-71c3-4b84-811d-eb77b29ac72d, please check neutron logs for more information. [ 664.986597] env[61906]: ERROR nova.compute.manager [ 664.986597] env[61906]: Traceback (most recent call last): [ 664.986597] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 664.986597] env[61906]: listener.cb(fileno) [ 664.986597] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 664.986597] env[61906]: result = function(*args, **kwargs) [ 664.986597] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 664.986597] env[61906]: return func(*args, **kwargs) [ 664.986597] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 664.986597] env[61906]: raise e [ 664.986597] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.986597] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 664.986597] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 664.986597] env[61906]: created_port_ids = self._update_ports_for_instance( [ 664.986597] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 664.986597] env[61906]: with excutils.save_and_reraise_exception(): [ 664.986597] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.986597] env[61906]: self.force_reraise() [ 664.986597] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.986597] env[61906]: raise self.value [ 664.986597] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 664.986597] env[61906]: updated_port = self._update_port( [ 664.986597] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.986597] env[61906]: _ensure_no_port_binding_failure(port) [ 664.986597] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.986597] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 664.987293] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 75dff989-71c3-4b84-811d-eb77b29ac72d, please check neutron logs for more information. [ 664.987293] env[61906]: Removing descriptor: 18 [ 665.036130] env[61906]: DEBUG nova.compute.manager [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 665.064447] env[61906]: DEBUG nova.virt.hardware [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 665.064696] env[61906]: DEBUG nova.virt.hardware [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 665.064893] env[61906]: DEBUG nova.virt.hardware [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 665.065075] env[61906]: DEBUG nova.virt.hardware [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 665.065550] env[61906]: DEBUG nova.virt.hardware [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 665.065550] env[61906]: DEBUG nova.virt.hardware [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 665.065550] env[61906]: DEBUG nova.virt.hardware [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 665.065788] env[61906]: DEBUG nova.virt.hardware [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 665.065903] env[61906]: DEBUG nova.virt.hardware [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 665.066049] env[61906]: DEBUG nova.virt.hardware [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 665.066224] env[61906]: DEBUG nova.virt.hardware [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 665.067120] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c7cd22-e90b-431d-b3de-4f72352a5539 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.075091] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34d3271d-2c0d-4098-91fe-f0d676be7fdd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.079731] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.124s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.080364] env[61906]: ERROR nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cd9987a6-8381-4928-a36b-b5d1193bf769, please check neutron logs for more information. [ 665.080364] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Traceback (most recent call last): [ 665.080364] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 665.080364] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] self.driver.spawn(context, instance, image_meta, [ 665.080364] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 665.080364] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 665.080364] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 665.080364] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] vm_ref = self.build_virtual_machine(instance, [ 665.080364] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 665.080364] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] vif_infos = vmwarevif.get_vif_info(self._session, [ 665.080364] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 665.080681] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] for vif in network_info: [ 665.080681] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 665.080681] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] return self._sync_wrapper(fn, *args, **kwargs) [ 665.080681] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 665.080681] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] self.wait() [ 665.080681] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 665.080681] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] self[:] = self._gt.wait() [ 665.080681] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 665.080681] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] return self._exit_event.wait() [ 665.080681] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 665.080681] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] current.throw(*self._exc) [ 665.080681] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.080681] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] result = function(*args, **kwargs) [ 665.080983] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 665.080983] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] return func(*args, **kwargs) [ 665.080983] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.080983] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] raise e [ 665.080983] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.080983] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] nwinfo = self.network_api.allocate_for_instance( [ 665.080983] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 665.080983] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] created_port_ids = self._update_ports_for_instance( [ 665.080983] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 665.080983] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] with excutils.save_and_reraise_exception(): [ 665.080983] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.080983] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] self.force_reraise() [ 665.080983] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.081332] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] raise self.value [ 665.081332] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 665.081332] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] updated_port = self._update_port( [ 665.081332] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.081332] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] _ensure_no_port_binding_failure(port) [ 665.081332] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.081332] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] raise exception.PortBindingFailed(port_id=port['id']) [ 665.081332] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] nova.exception.PortBindingFailed: Binding failed for port cd9987a6-8381-4928-a36b-b5d1193bf769, please check neutron logs for more information. [ 665.081332] env[61906]: ERROR nova.compute.manager [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] [ 665.081332] env[61906]: DEBUG nova.compute.utils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Binding failed for port cd9987a6-8381-4928-a36b-b5d1193bf769, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 665.082572] env[61906]: DEBUG oslo_concurrency.lockutils [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.395s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.083645] env[61906]: INFO nova.compute.claims [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 665.086561] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Build of instance cf331f9d-02ac-45d4-b938-0ed13c63e87b was re-scheduled: Binding failed for port cd9987a6-8381-4928-a36b-b5d1193bf769, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 665.087020] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 665.087239] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquiring lock "refresh_cache-cf331f9d-02ac-45d4-b938-0ed13c63e87b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.087385] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquired lock "refresh_cache-cf331f9d-02ac-45d4-b938-0ed13c63e87b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.087541] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 665.096763] env[61906]: ERROR nova.compute.manager [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 75dff989-71c3-4b84-811d-eb77b29ac72d, please check neutron logs for more information. [ 665.096763] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Traceback (most recent call last): [ 665.096763] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 665.096763] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] yield resources [ 665.096763] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 665.096763] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] self.driver.spawn(context, instance, image_meta, [ 665.096763] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 665.096763] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] self._vmops.spawn(context, instance, image_meta, injected_files, [ 665.096763] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 665.096763] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] vm_ref = self.build_virtual_machine(instance, [ 665.096763] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 665.097119] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] vif_infos = vmwarevif.get_vif_info(self._session, [ 665.097119] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 665.097119] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] for vif in network_info: [ 665.097119] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 665.097119] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] return self._sync_wrapper(fn, *args, **kwargs) [ 665.097119] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 665.097119] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] self.wait() [ 665.097119] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 665.097119] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] self[:] = self._gt.wait() [ 665.097119] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 665.097119] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] return self._exit_event.wait() [ 665.097119] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 665.097119] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] current.throw(*self._exc) [ 665.097495] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.097495] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] result = function(*args, **kwargs) [ 665.097495] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 665.097495] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] return func(*args, **kwargs) [ 665.097495] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.097495] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] raise e [ 665.097495] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.097495] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] nwinfo = self.network_api.allocate_for_instance( [ 665.097495] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 665.097495] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] created_port_ids = self._update_ports_for_instance( [ 665.097495] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 665.097495] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] with excutils.save_and_reraise_exception(): [ 665.097495] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.097885] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] self.force_reraise() [ 665.097885] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.097885] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] raise self.value [ 665.097885] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 665.097885] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] updated_port = self._update_port( [ 665.097885] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.097885] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] _ensure_no_port_binding_failure(port) [ 665.097885] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.097885] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] raise exception.PortBindingFailed(port_id=port['id']) [ 665.097885] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] nova.exception.PortBindingFailed: Binding failed for port 75dff989-71c3-4b84-811d-eb77b29ac72d, please check neutron logs for more information. [ 665.097885] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] [ 665.097885] env[61906]: INFO nova.compute.manager [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Terminating instance [ 665.100395] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Acquiring lock "refresh_cache-b6f46e45-742b-4801-86f0-fd8eea3611da" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.100567] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Acquired lock "refresh_cache-b6f46e45-742b-4801-86f0-fd8eea3611da" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.100733] env[61906]: DEBUG nova.network.neutron [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 665.623791] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.632150] env[61906]: DEBUG nova.network.neutron [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.723242] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.760820] env[61906]: DEBUG nova.network.neutron [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.227686] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Releasing lock "refresh_cache-cf331f9d-02ac-45d4-b938-0ed13c63e87b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.227686] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 666.227686] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 666.227686] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 666.250423] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 666.265115] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Releasing lock "refresh_cache-b6f46e45-742b-4801-86f0-fd8eea3611da" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.265530] env[61906]: DEBUG nova.compute.manager [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 666.265720] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 666.266461] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7e603dd2-a710-40d4-a644-e8204f5549e1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.284693] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69c8a0ac-d01f-48c1-81fb-2b289bc045b3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.310733] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b6f46e45-742b-4801-86f0-fd8eea3611da could not be found. [ 666.311130] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 666.311210] env[61906]: INFO nova.compute.manager [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Took 0.05 seconds to destroy the instance on the hypervisor. [ 666.311423] env[61906]: DEBUG oslo.service.loopingcall [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 666.313877] env[61906]: DEBUG nova.compute.manager [-] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 666.313984] env[61906]: DEBUG nova.network.neutron [-] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 666.337838] env[61906]: DEBUG nova.network.neutron [-] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 666.579351] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a9d468a-3d6a-4bc7-9769-69a9e19abd98 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.587131] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5cd5217-9d75-4231-bab7-4ef2060dc4b2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.618551] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d9c9ee0-4c15-40e7-9fbe-f2c89194eca4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.627084] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5415173-022a-4309-b9c4-db51c4e1d773 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.640238] env[61906]: DEBUG nova.compute.provider_tree [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 666.698811] env[61906]: DEBUG nova.compute.manager [req-6bcf826a-8e3e-4fe4-9de6-b1e1d37d6a7e req-12fa1779-faa9-4e09-8e19-c185450f5c48 service nova] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Received event network-changed-75dff989-71c3-4b84-811d-eb77b29ac72d {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 666.699032] env[61906]: DEBUG nova.compute.manager [req-6bcf826a-8e3e-4fe4-9de6-b1e1d37d6a7e req-12fa1779-faa9-4e09-8e19-c185450f5c48 service nova] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Refreshing instance network info cache due to event network-changed-75dff989-71c3-4b84-811d-eb77b29ac72d. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 666.699318] env[61906]: DEBUG oslo_concurrency.lockutils [req-6bcf826a-8e3e-4fe4-9de6-b1e1d37d6a7e req-12fa1779-faa9-4e09-8e19-c185450f5c48 service nova] Acquiring lock "refresh_cache-b6f46e45-742b-4801-86f0-fd8eea3611da" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.699456] env[61906]: DEBUG oslo_concurrency.lockutils [req-6bcf826a-8e3e-4fe4-9de6-b1e1d37d6a7e req-12fa1779-faa9-4e09-8e19-c185450f5c48 service nova] Acquired lock "refresh_cache-b6f46e45-742b-4801-86f0-fd8eea3611da" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.699591] env[61906]: DEBUG nova.network.neutron [req-6bcf826a-8e3e-4fe4-9de6-b1e1d37d6a7e req-12fa1779-faa9-4e09-8e19-c185450f5c48 service nova] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Refreshing network info cache for port 75dff989-71c3-4b84-811d-eb77b29ac72d {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 666.755914] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.842570] env[61906]: DEBUG nova.network.neutron [-] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.147019] env[61906]: DEBUG nova.scheduler.client.report [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 667.227064] env[61906]: DEBUG nova.network.neutron [req-6bcf826a-8e3e-4fe4-9de6-b1e1d37d6a7e req-12fa1779-faa9-4e09-8e19-c185450f5c48 service nova] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 667.260112] env[61906]: INFO nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Took 1.03 seconds to deallocate network for instance. [ 667.321315] env[61906]: DEBUG nova.network.neutron [req-6bcf826a-8e3e-4fe4-9de6-b1e1d37d6a7e req-12fa1779-faa9-4e09-8e19-c185450f5c48 service nova] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.346174] env[61906]: INFO nova.compute.manager [-] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Took 1.03 seconds to deallocate network for instance. [ 667.348384] env[61906]: DEBUG nova.compute.claims [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 667.348539] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.650709] env[61906]: DEBUG oslo_concurrency.lockutils [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.568s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.651266] env[61906]: DEBUG nova.compute.manager [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 667.655630] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.664s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.655630] env[61906]: INFO nova.compute.claims [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 667.823894] env[61906]: DEBUG oslo_concurrency.lockutils [req-6bcf826a-8e3e-4fe4-9de6-b1e1d37d6a7e req-12fa1779-faa9-4e09-8e19-c185450f5c48 service nova] Releasing lock "refresh_cache-b6f46e45-742b-4801-86f0-fd8eea3611da" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.824641] env[61906]: DEBUG nova.compute.manager [req-6bcf826a-8e3e-4fe4-9de6-b1e1d37d6a7e req-12fa1779-faa9-4e09-8e19-c185450f5c48 service nova] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Received event network-vif-deleted-75dff989-71c3-4b84-811d-eb77b29ac72d {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 668.163171] env[61906]: DEBUG nova.compute.utils [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 668.167219] env[61906]: DEBUG nova.compute.manager [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 668.167388] env[61906]: DEBUG nova.network.neutron [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 668.212392] env[61906]: DEBUG nova.policy [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd70d8c5a692d4a0e918516b9d5fb0983', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '838af940c9d54a6f93e8f18cbdca4e7f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 668.310292] env[61906]: INFO nova.scheduler.client.report [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Deleted allocations for instance cf331f9d-02ac-45d4-b938-0ed13c63e87b [ 668.544630] env[61906]: DEBUG nova.network.neutron [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Successfully created port: ead9d36e-fa05-4276-b0f1-2e9e005fe203 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 668.670963] env[61906]: DEBUG nova.compute.manager [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 668.821408] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "cf331f9d-02ac-45d4-b938-0ed13c63e87b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 95.569s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.823308] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bce7d229-d0b5-430a-b2e3-2625c02a7950 tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "cf331f9d-02ac-45d4-b938-0ed13c63e87b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 8.291s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.823542] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bce7d229-d0b5-430a-b2e3-2625c02a7950 tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquiring lock "cf331f9d-02ac-45d4-b938-0ed13c63e87b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.823743] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bce7d229-d0b5-430a-b2e3-2625c02a7950 tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "cf331f9d-02ac-45d4-b938-0ed13c63e87b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.823905] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bce7d229-d0b5-430a-b2e3-2625c02a7950 tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "cf331f9d-02ac-45d4-b938-0ed13c63e87b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.825919] env[61906]: INFO nova.compute.manager [None req-bce7d229-d0b5-430a-b2e3-2625c02a7950 tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Terminating instance [ 668.827463] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bce7d229-d0b5-430a-b2e3-2625c02a7950 tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquiring lock "refresh_cache-cf331f9d-02ac-45d4-b938-0ed13c63e87b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.828143] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bce7d229-d0b5-430a-b2e3-2625c02a7950 tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquired lock "refresh_cache-cf331f9d-02ac-45d4-b938-0ed13c63e87b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.828143] env[61906]: DEBUG nova.network.neutron [None req-bce7d229-d0b5-430a-b2e3-2625c02a7950 tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 669.136064] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-043bbcc9-be05-4362-a924-90bb0db2a11b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.145093] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cea6e1d-f7ef-4bf0-8d42-da2036a089b9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.174726] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa260969-8a27-459b-97c5-72187b623df4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.187170] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8717e4d-77c5-41ae-a1af-fec1d2bbd1ec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.201275] env[61906]: DEBUG nova.compute.provider_tree [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 669.326286] env[61906]: DEBUG nova.compute.manager [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 669.349037] env[61906]: DEBUG nova.network.neutron [None req-bce7d229-d0b5-430a-b2e3-2625c02a7950 tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.456475] env[61906]: DEBUG nova.network.neutron [None req-bce7d229-d0b5-430a-b2e3-2625c02a7950 tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.615915] env[61906]: DEBUG nova.compute.manager [req-d6e37a47-f59a-44f8-916f-df6deeab99a4 req-28a4996c-695e-4d2c-9915-a4ca5399d158 service nova] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Received event network-changed-ead9d36e-fa05-4276-b0f1-2e9e005fe203 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 669.616099] env[61906]: DEBUG nova.compute.manager [req-d6e37a47-f59a-44f8-916f-df6deeab99a4 req-28a4996c-695e-4d2c-9915-a4ca5399d158 service nova] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Refreshing instance network info cache due to event network-changed-ead9d36e-fa05-4276-b0f1-2e9e005fe203. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 669.616319] env[61906]: DEBUG oslo_concurrency.lockutils [req-d6e37a47-f59a-44f8-916f-df6deeab99a4 req-28a4996c-695e-4d2c-9915-a4ca5399d158 service nova] Acquiring lock "refresh_cache-0a1b97fe-42fb-47bd-bba2-aeb75c0710fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.616464] env[61906]: DEBUG oslo_concurrency.lockutils [req-d6e37a47-f59a-44f8-916f-df6deeab99a4 req-28a4996c-695e-4d2c-9915-a4ca5399d158 service nova] Acquired lock "refresh_cache-0a1b97fe-42fb-47bd-bba2-aeb75c0710fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.616623] env[61906]: DEBUG nova.network.neutron [req-d6e37a47-f59a-44f8-916f-df6deeab99a4 req-28a4996c-695e-4d2c-9915-a4ca5399d158 service nova] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Refreshing network info cache for port ead9d36e-fa05-4276-b0f1-2e9e005fe203 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 669.682915] env[61906]: DEBUG nova.compute.manager [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 669.706374] env[61906]: DEBUG nova.scheduler.client.report [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 669.722020] env[61906]: DEBUG nova.virt.hardware [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 669.722020] env[61906]: DEBUG nova.virt.hardware [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 669.722020] env[61906]: DEBUG nova.virt.hardware [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 669.722236] env[61906]: DEBUG nova.virt.hardware [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 669.722236] env[61906]: DEBUG nova.virt.hardware [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 669.722476] env[61906]: DEBUG nova.virt.hardware [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 669.722817] env[61906]: DEBUG nova.virt.hardware [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 669.723108] env[61906]: DEBUG nova.virt.hardware [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 669.723383] env[61906]: DEBUG nova.virt.hardware [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 669.723659] env[61906]: DEBUG nova.virt.hardware [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 669.723942] env[61906]: DEBUG nova.virt.hardware [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 669.726019] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b45968e-3d3a-432e-923c-98fed4c3d032 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.736800] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7da29d2e-aa85-4293-a1bd-ddbf7ebfd081 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.779883] env[61906]: ERROR nova.compute.manager [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ead9d36e-fa05-4276-b0f1-2e9e005fe203, please check neutron logs for more information. [ 669.779883] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 669.779883] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.779883] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 669.779883] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 669.779883] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 669.779883] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 669.779883] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 669.779883] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.779883] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 669.779883] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.779883] env[61906]: ERROR nova.compute.manager raise self.value [ 669.779883] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 669.779883] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 669.779883] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.779883] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 669.780411] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.780411] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 669.780411] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ead9d36e-fa05-4276-b0f1-2e9e005fe203, please check neutron logs for more information. [ 669.780411] env[61906]: ERROR nova.compute.manager [ 669.780411] env[61906]: Traceback (most recent call last): [ 669.780411] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 669.780411] env[61906]: listener.cb(fileno) [ 669.780411] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.780411] env[61906]: result = function(*args, **kwargs) [ 669.780411] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 669.780411] env[61906]: return func(*args, **kwargs) [ 669.780411] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.780411] env[61906]: raise e [ 669.780411] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.780411] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 669.780411] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 669.780411] env[61906]: created_port_ids = self._update_ports_for_instance( [ 669.780411] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 669.780411] env[61906]: with excutils.save_and_reraise_exception(): [ 669.780411] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.780411] env[61906]: self.force_reraise() [ 669.780411] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.780411] env[61906]: raise self.value [ 669.780411] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 669.780411] env[61906]: updated_port = self._update_port( [ 669.780411] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.780411] env[61906]: _ensure_no_port_binding_failure(port) [ 669.780411] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.780411] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 669.781310] env[61906]: nova.exception.PortBindingFailed: Binding failed for port ead9d36e-fa05-4276-b0f1-2e9e005fe203, please check neutron logs for more information. [ 669.781310] env[61906]: Removing descriptor: 16 [ 669.781310] env[61906]: ERROR nova.compute.manager [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ead9d36e-fa05-4276-b0f1-2e9e005fe203, please check neutron logs for more information. [ 669.781310] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Traceback (most recent call last): [ 669.781310] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 669.781310] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] yield resources [ 669.781310] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 669.781310] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] self.driver.spawn(context, instance, image_meta, [ 669.781310] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 669.781310] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 669.781310] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 669.781310] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] vm_ref = self.build_virtual_machine(instance, [ 669.781599] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 669.781599] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] vif_infos = vmwarevif.get_vif_info(self._session, [ 669.781599] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 669.781599] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] for vif in network_info: [ 669.781599] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 669.781599] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] return self._sync_wrapper(fn, *args, **kwargs) [ 669.781599] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 669.781599] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] self.wait() [ 669.781599] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 669.781599] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] self[:] = self._gt.wait() [ 669.781599] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 669.781599] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] return self._exit_event.wait() [ 669.781599] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 669.781966] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] result = hub.switch() [ 669.781966] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 669.781966] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] return self.greenlet.switch() [ 669.781966] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.781966] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] result = function(*args, **kwargs) [ 669.781966] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 669.781966] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] return func(*args, **kwargs) [ 669.781966] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.781966] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] raise e [ 669.781966] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.781966] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] nwinfo = self.network_api.allocate_for_instance( [ 669.781966] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 669.781966] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] created_port_ids = self._update_ports_for_instance( [ 669.782346] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 669.782346] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] with excutils.save_and_reraise_exception(): [ 669.782346] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.782346] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] self.force_reraise() [ 669.782346] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.782346] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] raise self.value [ 669.782346] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 669.782346] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] updated_port = self._update_port( [ 669.782346] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.782346] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] _ensure_no_port_binding_failure(port) [ 669.782346] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.782346] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] raise exception.PortBindingFailed(port_id=port['id']) [ 669.782735] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] nova.exception.PortBindingFailed: Binding failed for port ead9d36e-fa05-4276-b0f1-2e9e005fe203, please check neutron logs for more information. [ 669.782735] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] [ 669.782735] env[61906]: INFO nova.compute.manager [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Terminating instance [ 669.782735] env[61906]: DEBUG oslo_concurrency.lockutils [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Acquiring lock "refresh_cache-0a1b97fe-42fb-47bd-bba2-aeb75c0710fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.851894] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.961146] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bce7d229-d0b5-430a-b2e3-2625c02a7950 tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Releasing lock "refresh_cache-cf331f9d-02ac-45d4-b938-0ed13c63e87b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 669.961635] env[61906]: DEBUG nova.compute.manager [None req-bce7d229-d0b5-430a-b2e3-2625c02a7950 tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 669.961834] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bce7d229-d0b5-430a-b2e3-2625c02a7950 tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 669.962157] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-60e820de-8fe2-4aae-9d02-87841b4f187f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.971022] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-222f144b-bdcd-4232-9014-7c23a70e9541 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.993971] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-bce7d229-d0b5-430a-b2e3-2625c02a7950 tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cf331f9d-02ac-45d4-b938-0ed13c63e87b could not be found. [ 669.994392] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bce7d229-d0b5-430a-b2e3-2625c02a7950 tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 669.994674] env[61906]: INFO nova.compute.manager [None req-bce7d229-d0b5-430a-b2e3-2625c02a7950 tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 669.995053] env[61906]: DEBUG oslo.service.loopingcall [None req-bce7d229-d0b5-430a-b2e3-2625c02a7950 tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 669.995828] env[61906]: DEBUG nova.compute.manager [-] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 669.996013] env[61906]: DEBUG nova.network.neutron [-] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 670.012712] env[61906]: DEBUG nova.network.neutron [-] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.138765] env[61906]: DEBUG nova.network.neutron [req-d6e37a47-f59a-44f8-916f-df6deeab99a4 req-28a4996c-695e-4d2c-9915-a4ca5399d158 service nova] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.214518] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.560s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.216128] env[61906]: DEBUG nova.compute.manager [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 670.217687] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.226s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.221019] env[61906]: INFO nova.compute.claims [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 670.227033] env[61906]: DEBUG nova.network.neutron [req-d6e37a47-f59a-44f8-916f-df6deeab99a4 req-28a4996c-695e-4d2c-9915-a4ca5399d158 service nova] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.517654] env[61906]: DEBUG nova.network.neutron [-] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.724622] env[61906]: DEBUG nova.compute.utils [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 670.731560] env[61906]: DEBUG nova.compute.manager [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 670.731560] env[61906]: DEBUG nova.network.neutron [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 670.733454] env[61906]: DEBUG oslo_concurrency.lockutils [req-d6e37a47-f59a-44f8-916f-df6deeab99a4 req-28a4996c-695e-4d2c-9915-a4ca5399d158 service nova] Releasing lock "refresh_cache-0a1b97fe-42fb-47bd-bba2-aeb75c0710fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.733998] env[61906]: DEBUG oslo_concurrency.lockutils [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Acquired lock "refresh_cache-0a1b97fe-42fb-47bd-bba2-aeb75c0710fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.734215] env[61906]: DEBUG nova.network.neutron [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 670.829238] env[61906]: DEBUG nova.policy [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d61e569122f4ea98265474e80a72397', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fa689bcacc1d4a7384d39e8824c45557', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 671.020080] env[61906]: INFO nova.compute.manager [-] [instance: cf331f9d-02ac-45d4-b938-0ed13c63e87b] Took 1.02 seconds to deallocate network for instance. [ 671.232427] env[61906]: DEBUG nova.compute.manager [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 671.275326] env[61906]: DEBUG nova.network.neutron [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.290685] env[61906]: DEBUG nova.network.neutron [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Successfully created port: 6479ad8a-7a5a-4806-a541-37a118722244 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 671.393518] env[61906]: DEBUG nova.network.neutron [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.643204] env[61906]: DEBUG nova.compute.manager [req-8fecd27a-ab80-47f9-b697-dc365f6e20dc req-6d257b18-50a8-49db-a0e6-35d754f8a6d0 service nova] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Received event network-vif-deleted-ead9d36e-fa05-4276-b0f1-2e9e005fe203 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 671.704352] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dd8c403-7493-49ae-9d05-bca30ae5b1ab {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.712440] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29778da2-8092-4678-8b52-811f50cb6053 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.748442] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0542732b-11cd-4c84-96fd-1062b3976ea6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.760386] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2ad6147-a2e5-4ab7-9542-a3fcc9cd7bcd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.776682] env[61906]: DEBUG nova.compute.provider_tree [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 671.898484] env[61906]: DEBUG oslo_concurrency.lockutils [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Releasing lock "refresh_cache-0a1b97fe-42fb-47bd-bba2-aeb75c0710fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.898484] env[61906]: DEBUG nova.compute.manager [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 671.898484] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 671.899101] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e94473c5-1b8c-470a-a6f4-277c69ddcbc3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.915749] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beee8286-c340-4c8d-8563-ddc729252220 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.946022] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb could not be found. [ 671.946022] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 671.946022] env[61906]: INFO nova.compute.manager [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Took 0.05 seconds to destroy the instance on the hypervisor. [ 671.946022] env[61906]: DEBUG oslo.service.loopingcall [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 671.946397] env[61906]: DEBUG nova.compute.manager [-] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 671.946574] env[61906]: DEBUG nova.network.neutron [-] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 671.966862] env[61906]: DEBUG nova.network.neutron [-] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.050969] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bce7d229-d0b5-430a-b2e3-2625c02a7950 tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "cf331f9d-02ac-45d4-b938-0ed13c63e87b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.228s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.254275] env[61906]: DEBUG nova.compute.manager [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 672.279322] env[61906]: DEBUG nova.scheduler.client.report [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 672.284528] env[61906]: DEBUG nova.virt.hardware [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 672.284762] env[61906]: DEBUG nova.virt.hardware [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 672.284917] env[61906]: DEBUG nova.virt.hardware [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 672.285114] env[61906]: DEBUG nova.virt.hardware [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 672.285263] env[61906]: DEBUG nova.virt.hardware [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 672.285409] env[61906]: DEBUG nova.virt.hardware [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 672.285607] env[61906]: DEBUG nova.virt.hardware [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 672.285796] env[61906]: DEBUG nova.virt.hardware [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 672.285913] env[61906]: DEBUG nova.virt.hardware [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 672.290171] env[61906]: DEBUG nova.virt.hardware [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 672.290452] env[61906]: DEBUG nova.virt.hardware [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 672.291341] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e4cc119-490d-42c6-a634-c43d4554a026 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.305789] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eee25fdf-8be1-4cd3-9548-25df5f234e35 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.470477] env[61906]: DEBUG nova.network.neutron [-] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.540075] env[61906]: ERROR nova.compute.manager [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6479ad8a-7a5a-4806-a541-37a118722244, please check neutron logs for more information. [ 672.540075] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 672.540075] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.540075] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 672.540075] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.540075] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 672.540075] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.540075] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 672.540075] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.540075] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 672.540075] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.540075] env[61906]: ERROR nova.compute.manager raise self.value [ 672.540075] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.540075] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 672.540075] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.540075] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 672.540694] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.540694] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 672.540694] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6479ad8a-7a5a-4806-a541-37a118722244, please check neutron logs for more information. [ 672.540694] env[61906]: ERROR nova.compute.manager [ 672.540694] env[61906]: Traceback (most recent call last): [ 672.540694] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 672.540694] env[61906]: listener.cb(fileno) [ 672.540694] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.540694] env[61906]: result = function(*args, **kwargs) [ 672.540694] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 672.540694] env[61906]: return func(*args, **kwargs) [ 672.540694] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 672.540694] env[61906]: raise e [ 672.540694] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.540694] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 672.540694] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.540694] env[61906]: created_port_ids = self._update_ports_for_instance( [ 672.540694] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.540694] env[61906]: with excutils.save_and_reraise_exception(): [ 672.540694] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.540694] env[61906]: self.force_reraise() [ 672.540694] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.540694] env[61906]: raise self.value [ 672.540694] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.540694] env[61906]: updated_port = self._update_port( [ 672.540694] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.540694] env[61906]: _ensure_no_port_binding_failure(port) [ 672.540694] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.540694] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 672.541444] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 6479ad8a-7a5a-4806-a541-37a118722244, please check neutron logs for more information. [ 672.541444] env[61906]: Removing descriptor: 16 [ 672.541444] env[61906]: ERROR nova.compute.manager [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6479ad8a-7a5a-4806-a541-37a118722244, please check neutron logs for more information. [ 672.541444] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Traceback (most recent call last): [ 672.541444] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 672.541444] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] yield resources [ 672.541444] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 672.541444] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] self.driver.spawn(context, instance, image_meta, [ 672.541444] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 672.541444] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] self._vmops.spawn(context, instance, image_meta, injected_files, [ 672.541444] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 672.541444] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] vm_ref = self.build_virtual_machine(instance, [ 672.541760] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 672.541760] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] vif_infos = vmwarevif.get_vif_info(self._session, [ 672.541760] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 672.541760] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] for vif in network_info: [ 672.541760] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 672.541760] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] return self._sync_wrapper(fn, *args, **kwargs) [ 672.541760] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 672.541760] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] self.wait() [ 672.541760] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 672.541760] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] self[:] = self._gt.wait() [ 672.541760] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 672.541760] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] return self._exit_event.wait() [ 672.541760] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 672.542098] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] result = hub.switch() [ 672.542098] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 672.542098] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] return self.greenlet.switch() [ 672.542098] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.542098] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] result = function(*args, **kwargs) [ 672.542098] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 672.542098] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] return func(*args, **kwargs) [ 672.542098] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 672.542098] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] raise e [ 672.542098] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.542098] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] nwinfo = self.network_api.allocate_for_instance( [ 672.542098] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.542098] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] created_port_ids = self._update_ports_for_instance( [ 672.542428] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.542428] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] with excutils.save_and_reraise_exception(): [ 672.542428] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.542428] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] self.force_reraise() [ 672.542428] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.542428] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] raise self.value [ 672.542428] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.542428] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] updated_port = self._update_port( [ 672.542428] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.542428] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] _ensure_no_port_binding_failure(port) [ 672.542428] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.542428] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] raise exception.PortBindingFailed(port_id=port['id']) [ 672.542729] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] nova.exception.PortBindingFailed: Binding failed for port 6479ad8a-7a5a-4806-a541-37a118722244, please check neutron logs for more information. [ 672.542729] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] [ 672.542729] env[61906]: INFO nova.compute.manager [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Terminating instance [ 672.544856] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Acquiring lock "refresh_cache-59887ef7-b2fc-486c-a368-cebde3313113" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.544856] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Acquired lock "refresh_cache-59887ef7-b2fc-486c-a368-cebde3313113" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.544856] env[61906]: DEBUG nova.network.neutron [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 672.798965] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.581s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.799549] env[61906]: DEBUG nova.compute.manager [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 672.802687] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.662s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.973580] env[61906]: INFO nova.compute.manager [-] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Took 1.03 seconds to deallocate network for instance. [ 672.975673] env[61906]: DEBUG nova.compute.claims [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 672.975673] env[61906]: DEBUG oslo_concurrency.lockutils [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.063451] env[61906]: DEBUG nova.network.neutron [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 673.146415] env[61906]: DEBUG nova.network.neutron [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.307496] env[61906]: DEBUG nova.compute.utils [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 673.308876] env[61906]: DEBUG nova.compute.manager [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 673.309057] env[61906]: DEBUG nova.network.neutron [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 673.359298] env[61906]: DEBUG nova.policy [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd70d8c5a692d4a0e918516b9d5fb0983', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '838af940c9d54a6f93e8f18cbdca4e7f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 673.649439] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Releasing lock "refresh_cache-59887ef7-b2fc-486c-a368-cebde3313113" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.649880] env[61906]: DEBUG nova.compute.manager [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 673.651341] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 673.651341] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f015e631-b063-4089-a5b4-91c45779f8b4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.662372] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a18c979e-31ae-48bf-97bd-a4f98fc4b414 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.698134] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 59887ef7-b2fc-486c-a368-cebde3313113 could not be found. [ 673.698444] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 673.698662] env[61906]: INFO nova.compute.manager [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Took 0.05 seconds to destroy the instance on the hypervisor. [ 673.700169] env[61906]: DEBUG oslo.service.loopingcall [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 673.700169] env[61906]: DEBUG nova.network.neutron [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Successfully created port: f3cfb5a1-a938-47ba-8043-bb3f72ab73fd {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 673.704650] env[61906]: DEBUG nova.compute.manager [-] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 673.704817] env[61906]: DEBUG nova.network.neutron [-] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 673.728745] env[61906]: DEBUG nova.network.neutron [-] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 673.810145] env[61906]: DEBUG nova.compute.manager [req-f825ca15-51b3-4624-9c26-a01fd2244c30 req-40207715-0b03-42cf-91ea-d809200451d8 service nova] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Received event network-changed-6479ad8a-7a5a-4806-a541-37a118722244 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 673.810681] env[61906]: DEBUG nova.compute.manager [req-f825ca15-51b3-4624-9c26-a01fd2244c30 req-40207715-0b03-42cf-91ea-d809200451d8 service nova] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Refreshing instance network info cache due to event network-changed-6479ad8a-7a5a-4806-a541-37a118722244. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 673.811091] env[61906]: DEBUG oslo_concurrency.lockutils [req-f825ca15-51b3-4624-9c26-a01fd2244c30 req-40207715-0b03-42cf-91ea-d809200451d8 service nova] Acquiring lock "refresh_cache-59887ef7-b2fc-486c-a368-cebde3313113" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.811359] env[61906]: DEBUG oslo_concurrency.lockutils [req-f825ca15-51b3-4624-9c26-a01fd2244c30 req-40207715-0b03-42cf-91ea-d809200451d8 service nova] Acquired lock "refresh_cache-59887ef7-b2fc-486c-a368-cebde3313113" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.812030] env[61906]: DEBUG nova.network.neutron [req-f825ca15-51b3-4624-9c26-a01fd2244c30 req-40207715-0b03-42cf-91ea-d809200451d8 service nova] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Refreshing network info cache for port 6479ad8a-7a5a-4806-a541-37a118722244 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 673.816938] env[61906]: DEBUG nova.compute.manager [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 673.884016] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7254d4ae-63be-4d72-b7b8-ece7b3a406d1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.891682] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ac577c4-0699-4b82-8ebf-9bcddb983cd0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.925320] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10e26b1b-565d-41ea-a00a-3982f5ab83a3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.933236] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58140b41-20eb-479c-9074-2dce212f9e29 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.948237] env[61906]: DEBUG nova.compute.provider_tree [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.231032] env[61906]: DEBUG nova.network.neutron [-] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.328027] env[61906]: DEBUG nova.network.neutron [req-f825ca15-51b3-4624-9c26-a01fd2244c30 req-40207715-0b03-42cf-91ea-d809200451d8 service nova] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.409023] env[61906]: DEBUG nova.network.neutron [req-f825ca15-51b3-4624-9c26-a01fd2244c30 req-40207715-0b03-42cf-91ea-d809200451d8 service nova] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.454507] env[61906]: DEBUG nova.scheduler.client.report [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 674.701727] env[61906]: ERROR nova.compute.manager [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f3cfb5a1-a938-47ba-8043-bb3f72ab73fd, please check neutron logs for more information. [ 674.701727] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 674.701727] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.701727] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 674.701727] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 674.701727] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 674.701727] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 674.701727] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 674.701727] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.701727] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 674.701727] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.701727] env[61906]: ERROR nova.compute.manager raise self.value [ 674.701727] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 674.701727] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 674.701727] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.701727] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 674.702192] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.702192] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 674.702192] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f3cfb5a1-a938-47ba-8043-bb3f72ab73fd, please check neutron logs for more information. [ 674.702192] env[61906]: ERROR nova.compute.manager [ 674.702192] env[61906]: Traceback (most recent call last): [ 674.702192] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 674.702192] env[61906]: listener.cb(fileno) [ 674.702192] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.702192] env[61906]: result = function(*args, **kwargs) [ 674.702192] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 674.702192] env[61906]: return func(*args, **kwargs) [ 674.702192] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 674.702192] env[61906]: raise e [ 674.702192] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.702192] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 674.702192] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 674.702192] env[61906]: created_port_ids = self._update_ports_for_instance( [ 674.702192] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 674.702192] env[61906]: with excutils.save_and_reraise_exception(): [ 674.702192] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.702192] env[61906]: self.force_reraise() [ 674.702192] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.702192] env[61906]: raise self.value [ 674.702192] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 674.702192] env[61906]: updated_port = self._update_port( [ 674.702192] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.702192] env[61906]: _ensure_no_port_binding_failure(port) [ 674.702192] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.702192] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 674.702929] env[61906]: nova.exception.PortBindingFailed: Binding failed for port f3cfb5a1-a938-47ba-8043-bb3f72ab73fd, please check neutron logs for more information. [ 674.702929] env[61906]: Removing descriptor: 16 [ 674.734399] env[61906]: INFO nova.compute.manager [-] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Took 1.03 seconds to deallocate network for instance. [ 674.736186] env[61906]: DEBUG nova.compute.claims [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 674.736366] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.827873] env[61906]: DEBUG nova.compute.manager [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 674.855924] env[61906]: DEBUG nova.virt.hardware [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 674.856197] env[61906]: DEBUG nova.virt.hardware [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 674.856359] env[61906]: DEBUG nova.virt.hardware [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 674.856542] env[61906]: DEBUG nova.virt.hardware [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 674.856689] env[61906]: DEBUG nova.virt.hardware [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 674.856833] env[61906]: DEBUG nova.virt.hardware [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 674.857059] env[61906]: DEBUG nova.virt.hardware [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 674.857233] env[61906]: DEBUG nova.virt.hardware [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 674.857448] env[61906]: DEBUG nova.virt.hardware [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 674.857564] env[61906]: DEBUG nova.virt.hardware [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 674.857733] env[61906]: DEBUG nova.virt.hardware [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 674.858600] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c8dd9a8-c6e9-4084-bb56-8e98d1e72d45 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.866813] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cd19989-9953-4504-b6e0-79abffff8555 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.879940] env[61906]: ERROR nova.compute.manager [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f3cfb5a1-a938-47ba-8043-bb3f72ab73fd, please check neutron logs for more information. [ 674.879940] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Traceback (most recent call last): [ 674.879940] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 674.879940] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] yield resources [ 674.879940] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 674.879940] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] self.driver.spawn(context, instance, image_meta, [ 674.879940] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 674.879940] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] self._vmops.spawn(context, instance, image_meta, injected_files, [ 674.879940] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 674.879940] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] vm_ref = self.build_virtual_machine(instance, [ 674.879940] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 674.880331] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] vif_infos = vmwarevif.get_vif_info(self._session, [ 674.880331] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 674.880331] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] for vif in network_info: [ 674.880331] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 674.880331] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] return self._sync_wrapper(fn, *args, **kwargs) [ 674.880331] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 674.880331] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] self.wait() [ 674.880331] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 674.880331] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] self[:] = self._gt.wait() [ 674.880331] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 674.880331] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] return self._exit_event.wait() [ 674.880331] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 674.880331] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] current.throw(*self._exc) [ 674.880706] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.880706] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] result = function(*args, **kwargs) [ 674.880706] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 674.880706] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] return func(*args, **kwargs) [ 674.880706] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 674.880706] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] raise e [ 674.880706] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.880706] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] nwinfo = self.network_api.allocate_for_instance( [ 674.880706] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 674.880706] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] created_port_ids = self._update_ports_for_instance( [ 674.880706] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 674.880706] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] with excutils.save_and_reraise_exception(): [ 674.880706] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.881088] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] self.force_reraise() [ 674.881088] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.881088] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] raise self.value [ 674.881088] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 674.881088] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] updated_port = self._update_port( [ 674.881088] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.881088] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] _ensure_no_port_binding_failure(port) [ 674.881088] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.881088] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] raise exception.PortBindingFailed(port_id=port['id']) [ 674.881088] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] nova.exception.PortBindingFailed: Binding failed for port f3cfb5a1-a938-47ba-8043-bb3f72ab73fd, please check neutron logs for more information. [ 674.881088] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] [ 674.881088] env[61906]: INFO nova.compute.manager [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Terminating instance [ 674.882357] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Acquiring lock "refresh_cache-ff6fa10c-c1cc-4e85-af43-c52f06e35974" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.882519] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Acquired lock "refresh_cache-ff6fa10c-c1cc-4e85-af43-c52f06e35974" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.882709] env[61906]: DEBUG nova.network.neutron [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 674.911169] env[61906]: DEBUG oslo_concurrency.lockutils [req-f825ca15-51b3-4624-9c26-a01fd2244c30 req-40207715-0b03-42cf-91ea-d809200451d8 service nova] Releasing lock "refresh_cache-59887ef7-b2fc-486c-a368-cebde3313113" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.911522] env[61906]: DEBUG nova.compute.manager [req-f825ca15-51b3-4624-9c26-a01fd2244c30 req-40207715-0b03-42cf-91ea-d809200451d8 service nova] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Received event network-vif-deleted-6479ad8a-7a5a-4806-a541-37a118722244 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 674.959396] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.157s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.959935] env[61906]: ERROR nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c68c78a3-7249-49ae-94de-57569a36a8d7, please check neutron logs for more information. [ 674.959935] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Traceback (most recent call last): [ 674.959935] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 674.959935] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] self.driver.spawn(context, instance, image_meta, [ 674.959935] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 674.959935] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 674.959935] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 674.959935] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] vm_ref = self.build_virtual_machine(instance, [ 674.959935] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 674.959935] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] vif_infos = vmwarevif.get_vif_info(self._session, [ 674.959935] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 674.960517] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] for vif in network_info: [ 674.960517] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 674.960517] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] return self._sync_wrapper(fn, *args, **kwargs) [ 674.960517] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 674.960517] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] self.wait() [ 674.960517] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 674.960517] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] self[:] = self._gt.wait() [ 674.960517] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 674.960517] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] return self._exit_event.wait() [ 674.960517] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 674.960517] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] current.throw(*self._exc) [ 674.960517] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.960517] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] result = function(*args, **kwargs) [ 674.960885] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 674.960885] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] return func(*args, **kwargs) [ 674.960885] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 674.960885] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] raise e [ 674.960885] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.960885] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] nwinfo = self.network_api.allocate_for_instance( [ 674.960885] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 674.960885] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] created_port_ids = self._update_ports_for_instance( [ 674.960885] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 674.960885] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] with excutils.save_and_reraise_exception(): [ 674.960885] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.960885] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] self.force_reraise() [ 674.960885] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.961352] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] raise self.value [ 674.961352] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 674.961352] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] updated_port = self._update_port( [ 674.961352] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.961352] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] _ensure_no_port_binding_failure(port) [ 674.961352] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.961352] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] raise exception.PortBindingFailed(port_id=port['id']) [ 674.961352] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] nova.exception.PortBindingFailed: Binding failed for port c68c78a3-7249-49ae-94de-57569a36a8d7, please check neutron logs for more information. [ 674.961352] env[61906]: ERROR nova.compute.manager [instance: 5eafc757-e09b-4751-ac50-3e918156832b] [ 674.961352] env[61906]: DEBUG nova.compute.utils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Binding failed for port c68c78a3-7249-49ae-94de-57569a36a8d7, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 674.961898] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.260s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.964716] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Build of instance 5eafc757-e09b-4751-ac50-3e918156832b was re-scheduled: Binding failed for port c68c78a3-7249-49ae-94de-57569a36a8d7, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 674.965164] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 674.965355] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquiring lock "refresh_cache-5eafc757-e09b-4751-ac50-3e918156832b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.965493] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquired lock "refresh_cache-5eafc757-e09b-4751-ac50-3e918156832b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.965656] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 675.400827] env[61906]: DEBUG nova.network.neutron [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 675.473568] env[61906]: DEBUG nova.network.neutron [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.487407] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 675.572808] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.819871] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cbfb387-4ec4-4f0e-82ac-6d9cd052c1aa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.828465] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36269f79-6300-40d7-a09c-662bacbe2854 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.860578] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87358fbd-299e-44cf-960c-36c2636e4121 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.864049] env[61906]: DEBUG nova.compute.manager [req-864df2e3-b0ef-4796-8df8-4b102fec18be req-aa428d0f-1590-4919-bb59-c11b2661e313 service nova] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Received event network-changed-f3cfb5a1-a938-47ba-8043-bb3f72ab73fd {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 675.864237] env[61906]: DEBUG nova.compute.manager [req-864df2e3-b0ef-4796-8df8-4b102fec18be req-aa428d0f-1590-4919-bb59-c11b2661e313 service nova] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Refreshing instance network info cache due to event network-changed-f3cfb5a1-a938-47ba-8043-bb3f72ab73fd. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 675.864426] env[61906]: DEBUG oslo_concurrency.lockutils [req-864df2e3-b0ef-4796-8df8-4b102fec18be req-aa428d0f-1590-4919-bb59-c11b2661e313 service nova] Acquiring lock "refresh_cache-ff6fa10c-c1cc-4e85-af43-c52f06e35974" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.870466] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb0b5852-d4db-4666-b921-e2118b55a9da {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.884086] env[61906]: DEBUG nova.compute.provider_tree [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 675.977034] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Releasing lock "refresh_cache-ff6fa10c-c1cc-4e85-af43-c52f06e35974" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.977482] env[61906]: DEBUG nova.compute.manager [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 675.977673] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 675.978312] env[61906]: DEBUG oslo_concurrency.lockutils [req-864df2e3-b0ef-4796-8df8-4b102fec18be req-aa428d0f-1590-4919-bb59-c11b2661e313 service nova] Acquired lock "refresh_cache-ff6fa10c-c1cc-4e85-af43-c52f06e35974" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.978498] env[61906]: DEBUG nova.network.neutron [req-864df2e3-b0ef-4796-8df8-4b102fec18be req-aa428d0f-1590-4919-bb59-c11b2661e313 service nova] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Refreshing network info cache for port f3cfb5a1-a938-47ba-8043-bb3f72ab73fd {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 675.979953] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8ffad167-b277-41f7-9074-bf7e5365550c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.988827] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cbdb896-c75b-46cb-8a17-41d7a950c3a5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.011317] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ff6fa10c-c1cc-4e85-af43-c52f06e35974 could not be found. [ 676.011317] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 676.011582] env[61906]: INFO nova.compute.manager [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Took 0.03 seconds to destroy the instance on the hypervisor. [ 676.011734] env[61906]: DEBUG oslo.service.loopingcall [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 676.011946] env[61906]: DEBUG nova.compute.manager [-] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 676.012524] env[61906]: DEBUG nova.network.neutron [-] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 676.026754] env[61906]: DEBUG nova.network.neutron [-] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.077769] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Releasing lock "refresh_cache-5eafc757-e09b-4751-ac50-3e918156832b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.078015] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 676.078229] env[61906]: DEBUG nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 676.078377] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 676.092073] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.387382] env[61906]: DEBUG nova.scheduler.client.report [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 676.497816] env[61906]: DEBUG nova.network.neutron [req-864df2e3-b0ef-4796-8df8-4b102fec18be req-aa428d0f-1590-4919-bb59-c11b2661e313 service nova] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.528950] env[61906]: DEBUG nova.network.neutron [-] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.577345] env[61906]: DEBUG nova.network.neutron [req-864df2e3-b0ef-4796-8df8-4b102fec18be req-aa428d0f-1590-4919-bb59-c11b2661e313 service nova] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.594017] env[61906]: DEBUG nova.network.neutron [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.892857] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.931s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.893503] env[61906]: ERROR nova.compute.manager [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port aa5a85af-fe8b-4622-97ee-2db24b015683, please check neutron logs for more information. [ 676.893503] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Traceback (most recent call last): [ 676.893503] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 676.893503] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] self.driver.spawn(context, instance, image_meta, [ 676.893503] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 676.893503] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 676.893503] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 676.893503] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] vm_ref = self.build_virtual_machine(instance, [ 676.893503] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 676.893503] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] vif_infos = vmwarevif.get_vif_info(self._session, [ 676.893503] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 676.893834] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] for vif in network_info: [ 676.893834] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 676.893834] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] return self._sync_wrapper(fn, *args, **kwargs) [ 676.893834] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 676.893834] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] self.wait() [ 676.893834] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 676.893834] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] self[:] = self._gt.wait() [ 676.893834] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 676.893834] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] return self._exit_event.wait() [ 676.893834] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 676.893834] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] current.throw(*self._exc) [ 676.893834] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 676.893834] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] result = function(*args, **kwargs) [ 676.894212] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 676.894212] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] return func(*args, **kwargs) [ 676.894212] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 676.894212] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] raise e [ 676.894212] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 676.894212] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] nwinfo = self.network_api.allocate_for_instance( [ 676.894212] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 676.894212] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] created_port_ids = self._update_ports_for_instance( [ 676.894212] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 676.894212] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] with excutils.save_and_reraise_exception(): [ 676.894212] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.894212] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] self.force_reraise() [ 676.894212] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.894579] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] raise self.value [ 676.894579] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 676.894579] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] updated_port = self._update_port( [ 676.894579] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.894579] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] _ensure_no_port_binding_failure(port) [ 676.894579] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.894579] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] raise exception.PortBindingFailed(port_id=port['id']) [ 676.894579] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] nova.exception.PortBindingFailed: Binding failed for port aa5a85af-fe8b-4622-97ee-2db24b015683, please check neutron logs for more information. [ 676.894579] env[61906]: ERROR nova.compute.manager [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] [ 676.894579] env[61906]: DEBUG nova.compute.utils [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Binding failed for port aa5a85af-fe8b-4622-97ee-2db24b015683, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 676.895562] env[61906]: DEBUG oslo_concurrency.lockutils [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.642s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.899048] env[61906]: INFO nova.compute.claims [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 676.901820] env[61906]: DEBUG nova.compute.manager [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Build of instance 3595e6c6-10f3-46ba-85c4-90e25de78c1b was re-scheduled: Binding failed for port aa5a85af-fe8b-4622-97ee-2db24b015683, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 676.902271] env[61906]: DEBUG nova.compute.manager [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 676.902494] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Acquiring lock "refresh_cache-3595e6c6-10f3-46ba-85c4-90e25de78c1b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.902643] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Acquired lock "refresh_cache-3595e6c6-10f3-46ba-85c4-90e25de78c1b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.902800] env[61906]: DEBUG nova.network.neutron [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 677.031380] env[61906]: INFO nova.compute.manager [-] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Took 1.02 seconds to deallocate network for instance. [ 677.034026] env[61906]: DEBUG nova.compute.claims [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 677.034026] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.080077] env[61906]: DEBUG oslo_concurrency.lockutils [req-864df2e3-b0ef-4796-8df8-4b102fec18be req-aa428d0f-1590-4919-bb59-c11b2661e313 service nova] Releasing lock "refresh_cache-ff6fa10c-c1cc-4e85-af43-c52f06e35974" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.080388] env[61906]: DEBUG nova.compute.manager [req-864df2e3-b0ef-4796-8df8-4b102fec18be req-aa428d0f-1590-4919-bb59-c11b2661e313 service nova] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Received event network-vif-deleted-f3cfb5a1-a938-47ba-8043-bb3f72ab73fd {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 677.096153] env[61906]: INFO nova.compute.manager [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Took 1.02 seconds to deallocate network for instance. [ 677.427665] env[61906]: DEBUG nova.network.neutron [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 677.484272] env[61906]: DEBUG nova.network.neutron [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.987793] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Releasing lock "refresh_cache-3595e6c6-10f3-46ba-85c4-90e25de78c1b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.988081] env[61906]: DEBUG nova.compute.manager [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 677.988355] env[61906]: DEBUG nova.compute.manager [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 677.988454] env[61906]: DEBUG nova.network.neutron [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 678.005532] env[61906]: DEBUG nova.network.neutron [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 678.122484] env[61906]: INFO nova.scheduler.client.report [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Deleted allocations for instance 5eafc757-e09b-4751-ac50-3e918156832b [ 678.266027] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b467b2d-8d08-41e1-b40f-483c3bcf00f8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.273372] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ecae236-c708-4a9e-917a-1cc578459360 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.303399] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd9198f-28d7-42a3-afff-615966a902b8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.310989] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9dcd3d0-9a1e-48ba-8e11-6ad8e6937c61 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.325576] env[61906]: DEBUG nova.compute.provider_tree [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.508021] env[61906]: DEBUG nova.network.neutron [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.633625] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fbcba66-5125-4ebb-8542-2e693082d6af tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "5eafc757-e09b-4751-ac50-3e918156832b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.355s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.637108] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fcc0788-eb9c-4b86-aa29-521cdd8b524d tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "5eafc757-e09b-4751-ac50-3e918156832b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 18.214s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.637522] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fcc0788-eb9c-4b86-aa29-521cdd8b524d tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquiring lock "5eafc757-e09b-4751-ac50-3e918156832b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.637593] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fcc0788-eb9c-4b86-aa29-521cdd8b524d tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "5eafc757-e09b-4751-ac50-3e918156832b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.637751] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fcc0788-eb9c-4b86-aa29-521cdd8b524d tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "5eafc757-e09b-4751-ac50-3e918156832b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.641926] env[61906]: INFO nova.compute.manager [None req-2fcc0788-eb9c-4b86-aa29-521cdd8b524d tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Terminating instance [ 678.642815] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fcc0788-eb9c-4b86-aa29-521cdd8b524d tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquiring lock "refresh_cache-5eafc757-e09b-4751-ac50-3e918156832b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.642970] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fcc0788-eb9c-4b86-aa29-521cdd8b524d tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Acquired lock "refresh_cache-5eafc757-e09b-4751-ac50-3e918156832b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.643153] env[61906]: DEBUG nova.network.neutron [None req-2fcc0788-eb9c-4b86-aa29-521cdd8b524d tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 678.828387] env[61906]: DEBUG nova.scheduler.client.report [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 679.011263] env[61906]: INFO nova.compute.manager [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] [instance: 3595e6c6-10f3-46ba-85c4-90e25de78c1b] Took 1.02 seconds to deallocate network for instance. [ 679.139986] env[61906]: DEBUG nova.compute.manager [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 679.177230] env[61906]: DEBUG nova.network.neutron [None req-2fcc0788-eb9c-4b86-aa29-521cdd8b524d tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 679.310092] env[61906]: DEBUG nova.network.neutron [None req-2fcc0788-eb9c-4b86-aa29-521cdd8b524d tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.337095] env[61906]: DEBUG oslo_concurrency.lockutils [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.439s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.337095] env[61906]: DEBUG nova.compute.manager [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 679.338053] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 19.035s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.661458] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.812589] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fcc0788-eb9c-4b86-aa29-521cdd8b524d tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Releasing lock "refresh_cache-5eafc757-e09b-4751-ac50-3e918156832b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.813025] env[61906]: DEBUG nova.compute.manager [None req-2fcc0788-eb9c-4b86-aa29-521cdd8b524d tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 679.813218] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2fcc0788-eb9c-4b86-aa29-521cdd8b524d tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 679.813517] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2476db0c-7ddc-4d22-9ca3-d40da5e91ff2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.822340] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf072851-62a6-498b-bfca-c606a3a31ad7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.845022] env[61906]: DEBUG nova.compute.utils [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 679.849274] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-2fcc0788-eb9c-4b86-aa29-521cdd8b524d tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5eafc757-e09b-4751-ac50-3e918156832b could not be found. [ 679.849464] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2fcc0788-eb9c-4b86-aa29-521cdd8b524d tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 679.849642] env[61906]: INFO nova.compute.manager [None req-2fcc0788-eb9c-4b86-aa29-521cdd8b524d tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 679.849871] env[61906]: DEBUG oslo.service.loopingcall [None req-2fcc0788-eb9c-4b86-aa29-521cdd8b524d tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 679.850464] env[61906]: DEBUG nova.compute.manager [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Not allocating networking since 'none' was specified. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 679.850649] env[61906]: DEBUG nova.compute.manager [-] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 679.850745] env[61906]: DEBUG nova.network.neutron [-] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 679.874410] env[61906]: DEBUG nova.network.neutron [-] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 680.044268] env[61906]: INFO nova.scheduler.client.report [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Deleted allocations for instance 3595e6c6-10f3-46ba-85c4-90e25de78c1b [ 680.351879] env[61906]: DEBUG nova.compute.manager [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 680.376804] env[61906]: DEBUG nova.network.neutron [-] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.379071] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 32e2f482-9ceb-44bc-8933-115876374b1c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 680.379208] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance b6f46e45-742b-4801-86f0-fd8eea3611da actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 680.379328] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 680.379441] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 59887ef7-b2fc-486c-a368-cebde3313113 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 680.379572] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance ff6fa10c-c1cc-4e85-af43-c52f06e35974 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 680.379697] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 410b0308-1dca-4d11-91ab-f6d89bd565e3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 680.554723] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1fd0ed5-8b76-4f5c-9577-6f3593340ec6 tempest-DeleteServersAdminTestJSON-1739304128 tempest-DeleteServersAdminTestJSON-1739304128-project-member] Lock "3595e6c6-10f3-46ba-85c4-90e25de78c1b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.508s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.881992] env[61906]: INFO nova.compute.manager [-] [instance: 5eafc757-e09b-4751-ac50-3e918156832b] Took 1.03 seconds to deallocate network for instance. [ 680.882732] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 2e8d82fd-5f70-408e-9d3b-74f3d81232d3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 681.064115] env[61906]: DEBUG nova.compute.manager [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 681.366965] env[61906]: DEBUG nova.compute.manager [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 681.386346] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance fd30cb23-f30f-4e20-ae67-36956b78e6ac has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 681.397911] env[61906]: DEBUG nova.virt.hardware [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 681.398616] env[61906]: DEBUG nova.virt.hardware [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 681.398616] env[61906]: DEBUG nova.virt.hardware [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 681.398616] env[61906]: DEBUG nova.virt.hardware [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 681.398775] env[61906]: DEBUG nova.virt.hardware [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 681.399495] env[61906]: DEBUG nova.virt.hardware [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 681.399495] env[61906]: DEBUG nova.virt.hardware [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 681.399614] env[61906]: DEBUG nova.virt.hardware [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 681.399769] env[61906]: DEBUG nova.virt.hardware [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 681.399938] env[61906]: DEBUG nova.virt.hardware [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 681.400269] env[61906]: DEBUG nova.virt.hardware [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 681.401381] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfa1eb24-61ac-431b-966e-6bdce96264be {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.413562] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21383aae-f780-4d94-866e-d8505c5aa06b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.430882] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Instance VIF info [] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 681.436853] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Creating folder: Project (295f9e2cf3a74df4803f73483b0c55d8). Parent ref: group-v284713. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 681.437153] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cbd79523-980e-4c53-bfe5-d80967504df3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.449499] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Created folder: Project (295f9e2cf3a74df4803f73483b0c55d8) in parent group-v284713. [ 681.451324] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Creating folder: Instances. Parent ref: group-v284726. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 681.451324] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-37c14385-cb38-43d6-aa44-cca310f31974 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.458681] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Created folder: Instances in parent group-v284726. [ 681.460089] env[61906]: DEBUG oslo.service.loopingcall [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 681.460089] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 681.460089] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0b9f78b4-6636-47dd-8e03-d0cbd4126209 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.476430] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 681.476430] env[61906]: value = "task-1333118" [ 681.476430] env[61906]: _type = "Task" [ 681.476430] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.484218] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333118, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.587944] env[61906]: DEBUG oslo_concurrency.lockutils [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 681.890675] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance fe14efda-3b92-4045-8eef-0e5e92a9c538 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 681.923234] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fcc0788-eb9c-4b86-aa29-521cdd8b524d tempest-ListServersNegativeTestJSON-1548168594 tempest-ListServersNegativeTestJSON-1548168594-project-member] Lock "5eafc757-e09b-4751-ac50-3e918156832b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.287s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.986747] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333118, 'name': CreateVM_Task, 'duration_secs': 0.256432} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.986877] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 681.987200] env[61906]: DEBUG oslo_concurrency.lockutils [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.987360] env[61906]: DEBUG oslo_concurrency.lockutils [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.987668] env[61906]: DEBUG oslo_concurrency.lockutils [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 681.987904] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e96f7647-a208-4abb-aef0-96263fb9c850 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.992202] env[61906]: DEBUG oslo_vmware.api [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Waiting for the task: (returnval){ [ 681.992202] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52782857-9bf2-aa39-0bdc-219c7fe0091d" [ 681.992202] env[61906]: _type = "Task" [ 681.992202] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.000571] env[61906]: DEBUG oslo_vmware.api [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52782857-9bf2-aa39-0bdc-219c7fe0091d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.395025] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 9e665b58-69e7-4ab9-b109-ab27725e66cc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 682.502758] env[61906]: DEBUG oslo_vmware.api [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52782857-9bf2-aa39-0bdc-219c7fe0091d, 'name': SearchDatastore_Task, 'duration_secs': 0.010303} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.503082] env[61906]: DEBUG oslo_concurrency.lockutils [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.503321] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 682.503551] env[61906]: DEBUG oslo_concurrency.lockutils [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.503698] env[61906]: DEBUG oslo_concurrency.lockutils [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.503874] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 682.504150] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6748fd54-9f0a-4615-b726-e3587c5ceaab {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.512162] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 682.513026] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 682.513066] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1dd425b-5396-4f70-8d69-c086cca22b8e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.518118] env[61906]: DEBUG oslo_vmware.api [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Waiting for the task: (returnval){ [ 682.518118] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52084e2c-3a22-9d42-425a-6c81668ded52" [ 682.518118] env[61906]: _type = "Task" [ 682.518118] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.525934] env[61906]: DEBUG oslo_vmware.api [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52084e2c-3a22-9d42-425a-6c81668ded52, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.898350] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 683.028047] env[61906]: DEBUG oslo_vmware.api [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52084e2c-3a22-9d42-425a-6c81668ded52, 'name': SearchDatastore_Task, 'duration_secs': 0.008188} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.028413] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d802149-27da-41cf-8c85-2d2c539d08ad {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.033228] env[61906]: DEBUG oslo_vmware.api [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Waiting for the task: (returnval){ [ 683.033228] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52cbce32-61ef-3513-5e59-219156d3fad6" [ 683.033228] env[61906]: _type = "Task" [ 683.033228] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.040532] env[61906]: DEBUG oslo_vmware.api [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52cbce32-61ef-3513-5e59-219156d3fad6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.402020] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 7b4a9ebc-47fb-4078-969a-07642284b916 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 683.547171] env[61906]: DEBUG oslo_vmware.api [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52cbce32-61ef-3513-5e59-219156d3fad6, 'name': SearchDatastore_Task, 'duration_secs': 0.008262} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.547171] env[61906]: DEBUG oslo_concurrency.lockutils [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.547171] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 410b0308-1dca-4d11-91ab-f6d89bd565e3/410b0308-1dca-4d11-91ab-f6d89bd565e3.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 683.547171] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-62fc5964-9716-461c-9c4c-7b1d37b10f28 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.553896] env[61906]: DEBUG oslo_vmware.api [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Waiting for the task: (returnval){ [ 683.553896] env[61906]: value = "task-1333119" [ 683.553896] env[61906]: _type = "Task" [ 683.553896] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.560708] env[61906]: DEBUG oslo_vmware.api [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333119, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.905727] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 7c389914-47a6-4ed4-aad6-e8aeeefbed7c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 684.062562] env[61906]: DEBUG oslo_vmware.api [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333119, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471453} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.062824] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 410b0308-1dca-4d11-91ab-f6d89bd565e3/410b0308-1dca-4d11-91ab-f6d89bd565e3.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 684.063040] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 684.063286] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1b82ebd8-9435-4cb6-9b53-2ce93763dec3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.069951] env[61906]: DEBUG oslo_vmware.api [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Waiting for the task: (returnval){ [ 684.069951] env[61906]: value = "task-1333120" [ 684.069951] env[61906]: _type = "Task" [ 684.069951] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.077675] env[61906]: DEBUG oslo_vmware.api [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333120, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.409760] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance c15bc75b-87dc-4602-81ea-dbebce1b7763 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 684.579867] env[61906]: DEBUG oslo_vmware.api [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333120, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.188238} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.584215] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 684.585040] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c46b0f20-487d-48bb-83b8-546345ce98bf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.609065] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Reconfiguring VM instance instance-0000001d to attach disk [datastore2] 410b0308-1dca-4d11-91ab-f6d89bd565e3/410b0308-1dca-4d11-91ab-f6d89bd565e3.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 684.609413] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6af2a137-e062-4938-9c92-488f953e23cf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.633577] env[61906]: DEBUG oslo_vmware.api [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Waiting for the task: (returnval){ [ 684.633577] env[61906]: value = "task-1333121" [ 684.633577] env[61906]: _type = "Task" [ 684.633577] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.642546] env[61906]: DEBUG oslo_vmware.api [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333121, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.918023] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance c4b41c42-e01f-464a-9423-1c0fcca75586 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 685.143259] env[61906]: DEBUG oslo_vmware.api [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333121, 'name': ReconfigVM_Task, 'duration_secs': 0.290805} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.143517] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Reconfigured VM instance instance-0000001d to attach disk [datastore2] 410b0308-1dca-4d11-91ab-f6d89bd565e3/410b0308-1dca-4d11-91ab-f6d89bd565e3.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 685.144150] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-12f8f6fc-2cd7-4cc9-ad39-f2794f6861dd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.152588] env[61906]: DEBUG oslo_vmware.api [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Waiting for the task: (returnval){ [ 685.152588] env[61906]: value = "task-1333122" [ 685.152588] env[61906]: _type = "Task" [ 685.152588] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.160687] env[61906]: DEBUG oslo_vmware.api [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333122, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.421410] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance c9cb7335-8977-4ddf-9d0e-ff2166abf8e7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 685.662426] env[61906]: DEBUG oslo_vmware.api [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333122, 'name': Rename_Task, 'duration_secs': 0.144041} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.662774] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 685.663037] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8ba005c4-b169-43f1-9c77-3f19c3605101 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.669044] env[61906]: DEBUG oslo_vmware.api [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Waiting for the task: (returnval){ [ 685.669044] env[61906]: value = "task-1333123" [ 685.669044] env[61906]: _type = "Task" [ 685.669044] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.677074] env[61906]: DEBUG oslo_vmware.api [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333123, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.925931] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance bf056a94-9e9b-4296-b621-7e8a7338330b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 686.184018] env[61906]: DEBUG oslo_vmware.api [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333123, 'name': PowerOnVM_Task, 'duration_secs': 0.417733} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.184018] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 686.184018] env[61906]: INFO nova.compute.manager [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Took 4.82 seconds to spawn the instance on the hypervisor. [ 686.184018] env[61906]: DEBUG nova.compute.manager [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 686.184018] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba77ade6-9ac8-4eec-9dd0-220d20a6cd85 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.429409] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance adcd5674-ce1e-4b73-85bd-4344b5633c94 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 686.703293] env[61906]: INFO nova.compute.manager [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Took 28.47 seconds to build instance. [ 686.932614] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance b071a9b5-3416-4898-9ae2-ef62186c1843 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 687.205106] env[61906]: DEBUG oslo_concurrency.lockutils [None req-acaea767-0406-4d6f-b420-50fe3e8783b9 tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Lock "410b0308-1dca-4d11-91ab-f6d89bd565e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.179s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.389725] env[61906]: DEBUG oslo_concurrency.lockutils [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Acquiring lock "df9c9d62-8dc5-452c-85ed-846dcbd76014" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.389990] env[61906]: DEBUG oslo_concurrency.lockutils [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Lock "df9c9d62-8dc5-452c-85ed-846dcbd76014" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.436180] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance a07e49a5-e20a-49df-9326-a2e819934d15 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 687.504567] env[61906]: INFO nova.compute.manager [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Rebuilding instance [ 687.548546] env[61906]: DEBUG nova.compute.manager [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 687.549440] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe373b2e-2283-49b5-8c78-886bd49cee4c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.707834] env[61906]: DEBUG nova.compute.manager [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 687.939764] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 58b454d7-5370-499a-93dc-8613ee997b4d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 688.060609] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 688.060925] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a9311ffd-3770-4b49-8881-69f601a315d3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.068182] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Waiting for the task: (returnval){ [ 688.068182] env[61906]: value = "task-1333124" [ 688.068182] env[61906]: _type = "Task" [ 688.068182] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.077299] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333124, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.230124] env[61906]: DEBUG oslo_concurrency.lockutils [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.443028] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 3df44b70-ee72-40d7-9129-b5e94e121b0a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 688.577747] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333124, 'name': PowerOffVM_Task, 'duration_secs': 0.184312} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.578049] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 688.578276] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 688.579023] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b84c3e44-f748-483d-9e57-a55d439fe82b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.585470] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 688.585682] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b98c28a2-85f4-443a-b8a9-c352fab4346a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.612229] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 688.612463] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 688.612641] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Deleting the datastore file [datastore2] 410b0308-1dca-4d11-91ab-f6d89bd565e3 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 688.612962] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-99237adf-dc7a-42cd-a3a2-bb2286defd1e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.619673] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Waiting for the task: (returnval){ [ 688.619673] env[61906]: value = "task-1333126" [ 688.619673] env[61906]: _type = "Task" [ 688.619673] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.626689] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333126, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.945981] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance a048328a-d4a8-420e-9cc8-047254d7d1bf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 689.130176] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333126, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.093184} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.130478] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 689.130710] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 689.130890] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 689.448651] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 689.951404] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 830e7d50-5870-4e1f-8f2f-26840ba06d33 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 690.166096] env[61906]: DEBUG nova.virt.hardware [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 690.166096] env[61906]: DEBUG nova.virt.hardware [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 690.166096] env[61906]: DEBUG nova.virt.hardware [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 690.166096] env[61906]: DEBUG nova.virt.hardware [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 690.166400] env[61906]: DEBUG nova.virt.hardware [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 690.166400] env[61906]: DEBUG nova.virt.hardware [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 690.166400] env[61906]: DEBUG nova.virt.hardware [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 690.166400] env[61906]: DEBUG nova.virt.hardware [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 690.166579] env[61906]: DEBUG nova.virt.hardware [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 690.166579] env[61906]: DEBUG nova.virt.hardware [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 690.166844] env[61906]: DEBUG nova.virt.hardware [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 690.167607] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a871e37-57f5-4023-a13d-a093676be139 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.175460] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd7992d-41b0-4b61-91d1-bf00a0c7fc20 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.188587] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Instance VIF info [] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 690.194922] env[61906]: DEBUG oslo.service.loopingcall [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 690.195214] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 690.195482] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bfef9f66-3c5a-474b-b7b6-6317c6b0902e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.214071] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 690.214071] env[61906]: value = "task-1333127" [ 690.214071] env[61906]: _type = "Task" [ 690.214071] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.221426] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333127, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.455236] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 5ab321ec-655b-4f68-91aa-9b3db8c692f4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 690.724901] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333127, 'name': CreateVM_Task, 'duration_secs': 0.350802} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.725087] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 690.725492] env[61906]: DEBUG oslo_concurrency.lockutils [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.725651] env[61906]: DEBUG oslo_concurrency.lockutils [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.725954] env[61906]: DEBUG oslo_concurrency.lockutils [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 690.726207] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3efe45a1-daf9-412f-be62-38940d7d5291 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.730643] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Waiting for the task: (returnval){ [ 690.730643] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52b8c54f-9ffe-96b8-ac0f-02c5da03d755" [ 690.730643] env[61906]: _type = "Task" [ 690.730643] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.738350] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52b8c54f-9ffe-96b8-ac0f-02c5da03d755, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.965029] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 7ad136a2-d850-4b68-af0f-73badf65426c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 691.240983] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52b8c54f-9ffe-96b8-ac0f-02c5da03d755, 'name': SearchDatastore_Task, 'duration_secs': 0.012632} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.241365] env[61906]: DEBUG oslo_concurrency.lockutils [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.241605] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 691.241831] env[61906]: DEBUG oslo_concurrency.lockutils [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.241975] env[61906]: DEBUG oslo_concurrency.lockutils [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.242206] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 691.242421] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2fab5696-f717-4454-8f15-e6bfbeda348d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.251114] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 691.251307] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 691.252015] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ff2bd92-9885-4feb-93bf-3a691cc3c769 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.257132] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Waiting for the task: (returnval){ [ 691.257132] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52309215-e0c0-7531-f1c1-dfbcbcd1585d" [ 691.257132] env[61906]: _type = "Task" [ 691.257132] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.264638] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52309215-e0c0-7531-f1c1-dfbcbcd1585d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.467859] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 7e452659-0a5e-4740-b7ed-1fd7b1917dac has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 691.768282] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52309215-e0c0-7531-f1c1-dfbcbcd1585d, 'name': SearchDatastore_Task, 'duration_secs': 0.011719} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.769103] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b573ef89-a39e-49d4-9eb0-f837056b6bd0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.774291] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Waiting for the task: (returnval){ [ 691.774291] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52f5fb9c-b9c6-798f-b6e2-f781c21b39d3" [ 691.774291] env[61906]: _type = "Task" [ 691.774291] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.782142] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52f5fb9c-b9c6-798f-b6e2-f781c21b39d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.970529] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance b935c7ba-7716-443b-b5da-9ee03cef793a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 692.285839] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52f5fb9c-b9c6-798f-b6e2-f781c21b39d3, 'name': SearchDatastore_Task, 'duration_secs': 0.013477} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.286117] env[61906]: DEBUG oslo_concurrency.lockutils [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 692.286372] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 410b0308-1dca-4d11-91ab-f6d89bd565e3/410b0308-1dca-4d11-91ab-f6d89bd565e3.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 692.286630] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4c737e73-9f98-4d3d-b481-28a5c2a53c32 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.293496] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Waiting for the task: (returnval){ [ 692.293496] env[61906]: value = "task-1333128" [ 692.293496] env[61906]: _type = "Task" [ 692.293496] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.301973] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333128, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.473737] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 692.474072] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 692.474202] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 692.805559] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333128, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.880168] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61a61fb6-8390-483d-bce8-9aee7580d1cf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.887436] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ec9b33a-9c8e-4634-a9f0-06369a940a5e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.916737] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c940ef-3d71-405b-9d57-3d7120726e4b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.924380] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bd75c6e-a73b-40de-95e8-1ef2f6ed9f98 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.938061] env[61906]: DEBUG nova.compute.provider_tree [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 693.305347] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333128, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.547387} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.305624] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 410b0308-1dca-4d11-91ab-f6d89bd565e3/410b0308-1dca-4d11-91ab-f6d89bd565e3.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 693.305836] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 693.306098] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-421175af-a369-4eeb-97c8-82eba9375609 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.312282] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Waiting for the task: (returnval){ [ 693.312282] env[61906]: value = "task-1333129" [ 693.312282] env[61906]: _type = "Task" [ 693.312282] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.319978] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333129, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.441078] env[61906]: DEBUG nova.scheduler.client.report [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 693.823009] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333129, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.204724} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.823329] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 693.824137] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc60744d-db09-4339-8961-4a0df1d96a27 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.844447] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Reconfiguring VM instance instance-0000001d to attach disk [datastore2] 410b0308-1dca-4d11-91ab-f6d89bd565e3/410b0308-1dca-4d11-91ab-f6d89bd565e3.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 693.844693] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d5e283e-7a32-49c7-ae21-df6c222da884 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.863723] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Waiting for the task: (returnval){ [ 693.863723] env[61906]: value = "task-1333130" [ 693.863723] env[61906]: _type = "Task" [ 693.863723] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.871123] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333130, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.947131] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 693.947495] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 14.609s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.947808] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.605s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.949435] env[61906]: INFO nova.compute.claims [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 694.375143] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333130, 'name': ReconfigVM_Task, 'duration_secs': 0.426762} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.375613] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Reconfigured VM instance instance-0000001d to attach disk [datastore2] 410b0308-1dca-4d11-91ab-f6d89bd565e3/410b0308-1dca-4d11-91ab-f6d89bd565e3.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 694.376460] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b6097f25-54c4-4ca4-a062-a081530123a4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.383626] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Waiting for the task: (returnval){ [ 694.383626] env[61906]: value = "task-1333131" [ 694.383626] env[61906]: _type = "Task" [ 694.383626] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.392356] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333131, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.893708] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333131, 'name': Rename_Task, 'duration_secs': 0.148316} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.893996] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 694.894243] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2715a648-f004-4839-8659-ccc8a757fbd0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.899734] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Waiting for the task: (returnval){ [ 694.899734] env[61906]: value = "task-1333132" [ 694.899734] env[61906]: _type = "Task" [ 694.899734] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.906945] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333132, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.316785] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6274c73-e3f6-4b66-bc05-48ec9abf9b75 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.324412] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b82df59-9ae1-42d7-b2a6-ae6313b62dd6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.354051] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d069eba5-e1d3-47f2-9d06-b6bed68915fa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.361329] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6771ba4f-6190-4a7c-932f-d4ad254ba69e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.374076] env[61906]: DEBUG nova.compute.provider_tree [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 695.409741] env[61906]: DEBUG oslo_vmware.api [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333132, 'name': PowerOnVM_Task, 'duration_secs': 0.394969} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.410739] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 695.410739] env[61906]: DEBUG nova.compute.manager [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 695.411141] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c6104b-8ee9-4ffd-bdae-0f848e3e6947 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.879220] env[61906]: DEBUG nova.scheduler.client.report [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 695.929163] env[61906]: DEBUG oslo_concurrency.lockutils [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.263875] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Acquiring lock "410b0308-1dca-4d11-91ab-f6d89bd565e3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.264175] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Lock "410b0308-1dca-4d11-91ab-f6d89bd565e3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.264392] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Acquiring lock "410b0308-1dca-4d11-91ab-f6d89bd565e3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.264576] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Lock "410b0308-1dca-4d11-91ab-f6d89bd565e3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.264747] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Lock "410b0308-1dca-4d11-91ab-f6d89bd565e3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 696.266870] env[61906]: INFO nova.compute.manager [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Terminating instance [ 696.268817] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Acquiring lock "refresh_cache-410b0308-1dca-4d11-91ab-f6d89bd565e3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.268817] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Acquired lock "refresh_cache-410b0308-1dca-4d11-91ab-f6d89bd565e3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.268941] env[61906]: DEBUG nova.network.neutron [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 696.383838] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.436s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 696.384315] env[61906]: DEBUG nova.compute.manager [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 696.387131] env[61906]: DEBUG oslo_concurrency.lockutils [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.524s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.794803] env[61906]: DEBUG nova.network.neutron [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.854973] env[61906]: DEBUG nova.network.neutron [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.892697] env[61906]: DEBUG nova.compute.utils [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 696.897597] env[61906]: DEBUG nova.compute.manager [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 696.897775] env[61906]: DEBUG nova.network.neutron [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 696.945613] env[61906]: DEBUG nova.policy [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '68ac47440e284e4289b0aaf3bbeafdf5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fddbb54174c14b4782c89bcd3bd8f5e6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 697.253878] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cbbce50-74c7-4827-b9d6-b0d2539d21f0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.257718] env[61906]: DEBUG nova.network.neutron [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Successfully created port: 99389367-21dd-45ae-8f2e-c1002dc479dd {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 697.264411] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0923a75f-b4ee-473a-8997-202a72a9acc4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.294047] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-101d1aa4-dbcf-4941-a6fa-bbc2178bb74b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.301357] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5c999a2-807f-4fd3-855d-7ca4ec7544d3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.315703] env[61906]: DEBUG nova.compute.provider_tree [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 697.358047] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Releasing lock "refresh_cache-410b0308-1dca-4d11-91ab-f6d89bd565e3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.358499] env[61906]: DEBUG nova.compute.manager [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 697.358700] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 697.359539] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e18a23ca-d6ce-42d4-bc28-3db8da7507b5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.367674] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 697.367814] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-80dd1765-7e60-4364-9049-14d81ba50d2a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.374151] env[61906]: DEBUG oslo_vmware.api [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Waiting for the task: (returnval){ [ 697.374151] env[61906]: value = "task-1333133" [ 697.374151] env[61906]: _type = "Task" [ 697.374151] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.384886] env[61906]: DEBUG oslo_vmware.api [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333133, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.398419] env[61906]: DEBUG nova.compute.manager [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 697.819013] env[61906]: DEBUG nova.scheduler.client.report [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 697.887579] env[61906]: DEBUG oslo_vmware.api [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333133, 'name': PowerOffVM_Task, 'duration_secs': 0.110467} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.887832] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 697.888056] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 697.888258] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1050036d-35d1-4a31-b68b-58ea103e587a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.904934] env[61906]: INFO nova.virt.block_device [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Booting with volume 9e2bcfa1-0cda-4425-bb2c-8cb98dc2313a at /dev/sda [ 697.912732] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 697.912972] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 697.913175] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Deleting the datastore file [datastore2] 410b0308-1dca-4d11-91ab-f6d89bd565e3 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 697.913421] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d1956a98-fecb-4e92-9546-daef9fba56e4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.932930] env[61906]: DEBUG oslo_vmware.api [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Waiting for the task: (returnval){ [ 697.932930] env[61906]: value = "task-1333135" [ 697.932930] env[61906]: _type = "Task" [ 697.932930] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.942719] env[61906]: DEBUG oslo_vmware.api [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333135, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.964216] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d609fe19-6136-4cd4-9900-9b96813ce887 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.977616] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28466736-c43f-4387-9a69-efebf7841d82 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.998038] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fd272232-542d-4c24-9efe-e0e07ed611e0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.005522] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f5287b9-a545-43a7-a15c-b4cdd204f256 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.026929] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d80ce73e-d658-4fc7-918c-bdad7575f203 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.033992] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b7104f3-ef31-4369-8e13-c247f26f2b5f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.048086] env[61906]: DEBUG nova.virt.block_device [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Updating existing volume attachment record: a963770f-29f0-43a2-ba19-98ef24cc7f72 {{(pid=61906) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 698.071024] env[61906]: DEBUG nova.compute.manager [req-ebaee2ea-db16-4f99-9e06-43d8c932ad01 req-63672a65-cc54-4b6f-9482-c9d609ff4cdd service nova] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Received event network-changed-99389367-21dd-45ae-8f2e-c1002dc479dd {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 698.071024] env[61906]: DEBUG nova.compute.manager [req-ebaee2ea-db16-4f99-9e06-43d8c932ad01 req-63672a65-cc54-4b6f-9482-c9d609ff4cdd service nova] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Refreshing instance network info cache due to event network-changed-99389367-21dd-45ae-8f2e-c1002dc479dd. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 698.071024] env[61906]: DEBUG oslo_concurrency.lockutils [req-ebaee2ea-db16-4f99-9e06-43d8c932ad01 req-63672a65-cc54-4b6f-9482-c9d609ff4cdd service nova] Acquiring lock "refresh_cache-2e8d82fd-5f70-408e-9d3b-74f3d81232d3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.071024] env[61906]: DEBUG oslo_concurrency.lockutils [req-ebaee2ea-db16-4f99-9e06-43d8c932ad01 req-63672a65-cc54-4b6f-9482-c9d609ff4cdd service nova] Acquired lock "refresh_cache-2e8d82fd-5f70-408e-9d3b-74f3d81232d3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.071024] env[61906]: DEBUG nova.network.neutron [req-ebaee2ea-db16-4f99-9e06-43d8c932ad01 req-63672a65-cc54-4b6f-9482-c9d609ff4cdd service nova] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Refreshing network info cache for port 99389367-21dd-45ae-8f2e-c1002dc479dd {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 698.257765] env[61906]: ERROR nova.compute.manager [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 99389367-21dd-45ae-8f2e-c1002dc479dd, please check neutron logs for more information. [ 698.257765] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 698.257765] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 698.257765] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 698.257765] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 698.257765] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 698.257765] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 698.257765] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 698.257765] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 698.257765] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 698.257765] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 698.257765] env[61906]: ERROR nova.compute.manager raise self.value [ 698.257765] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 698.257765] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 698.257765] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 698.257765] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 698.258280] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 698.258280] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 698.258280] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 99389367-21dd-45ae-8f2e-c1002dc479dd, please check neutron logs for more information. [ 698.258280] env[61906]: ERROR nova.compute.manager [ 698.258280] env[61906]: Traceback (most recent call last): [ 698.258280] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 698.258280] env[61906]: listener.cb(fileno) [ 698.258280] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 698.258280] env[61906]: result = function(*args, **kwargs) [ 698.258280] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 698.258280] env[61906]: return func(*args, **kwargs) [ 698.258280] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 698.258280] env[61906]: raise e [ 698.258280] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 698.258280] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 698.258280] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 698.258280] env[61906]: created_port_ids = self._update_ports_for_instance( [ 698.258280] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 698.258280] env[61906]: with excutils.save_and_reraise_exception(): [ 698.258280] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 698.258280] env[61906]: self.force_reraise() [ 698.258280] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 698.258280] env[61906]: raise self.value [ 698.258280] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 698.258280] env[61906]: updated_port = self._update_port( [ 698.258280] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 698.258280] env[61906]: _ensure_no_port_binding_failure(port) [ 698.258280] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 698.258280] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 698.259497] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 99389367-21dd-45ae-8f2e-c1002dc479dd, please check neutron logs for more information. [ 698.259497] env[61906]: Removing descriptor: 18 [ 698.327146] env[61906]: DEBUG oslo_concurrency.lockutils [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.940s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.327776] env[61906]: ERROR nova.compute.manager [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9691890e-5d7f-4f1d-910a-a3d195288529, please check neutron logs for more information. [ 698.327776] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Traceback (most recent call last): [ 698.327776] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 698.327776] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] self.driver.spawn(context, instance, image_meta, [ 698.327776] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 698.327776] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 698.327776] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 698.327776] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] vm_ref = self.build_virtual_machine(instance, [ 698.327776] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 698.327776] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] vif_infos = vmwarevif.get_vif_info(self._session, [ 698.327776] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 698.328144] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] for vif in network_info: [ 698.328144] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 698.328144] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] return self._sync_wrapper(fn, *args, **kwargs) [ 698.328144] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 698.328144] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] self.wait() [ 698.328144] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 698.328144] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] self[:] = self._gt.wait() [ 698.328144] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 698.328144] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] return self._exit_event.wait() [ 698.328144] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 698.328144] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] result = hub.switch() [ 698.328144] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 698.328144] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] return self.greenlet.switch() [ 698.328511] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 698.328511] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] result = function(*args, **kwargs) [ 698.328511] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 698.328511] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] return func(*args, **kwargs) [ 698.328511] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 698.328511] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] raise e [ 698.328511] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 698.328511] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] nwinfo = self.network_api.allocate_for_instance( [ 698.328511] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 698.328511] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] created_port_ids = self._update_ports_for_instance( [ 698.328511] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 698.328511] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] with excutils.save_and_reraise_exception(): [ 698.328511] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 698.328906] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] self.force_reraise() [ 698.328906] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 698.328906] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] raise self.value [ 698.328906] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 698.328906] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] updated_port = self._update_port( [ 698.328906] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 698.328906] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] _ensure_no_port_binding_failure(port) [ 698.328906] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 698.328906] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] raise exception.PortBindingFailed(port_id=port['id']) [ 698.328906] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] nova.exception.PortBindingFailed: Binding failed for port 9691890e-5d7f-4f1d-910a-a3d195288529, please check neutron logs for more information. [ 698.328906] env[61906]: ERROR nova.compute.manager [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] [ 698.329233] env[61906]: DEBUG nova.compute.utils [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Binding failed for port 9691890e-5d7f-4f1d-910a-a3d195288529, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 698.329992] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.981s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.333478] env[61906]: DEBUG nova.compute.manager [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Build of instance 32e2f482-9ceb-44bc-8933-115876374b1c was re-scheduled: Binding failed for port 9691890e-5d7f-4f1d-910a-a3d195288529, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 698.333908] env[61906]: DEBUG nova.compute.manager [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 698.334141] env[61906]: DEBUG oslo_concurrency.lockutils [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Acquiring lock "refresh_cache-32e2f482-9ceb-44bc-8933-115876374b1c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.334650] env[61906]: DEBUG oslo_concurrency.lockutils [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Acquired lock "refresh_cache-32e2f482-9ceb-44bc-8933-115876374b1c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.334650] env[61906]: DEBUG nova.network.neutron [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 698.443161] env[61906]: DEBUG oslo_vmware.api [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Task: {'id': task-1333135, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.094676} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.443500] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 698.443698] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 698.443880] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 698.444064] env[61906]: INFO nova.compute.manager [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Took 1.09 seconds to destroy the instance on the hypervisor. [ 698.444319] env[61906]: DEBUG oslo.service.loopingcall [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 698.444499] env[61906]: DEBUG nova.compute.manager [-] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 698.444596] env[61906]: DEBUG nova.network.neutron [-] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 698.458941] env[61906]: DEBUG nova.network.neutron [-] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.589620] env[61906]: DEBUG nova.network.neutron [req-ebaee2ea-db16-4f99-9e06-43d8c932ad01 req-63672a65-cc54-4b6f-9482-c9d609ff4cdd service nova] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.665874] env[61906]: DEBUG nova.network.neutron [req-ebaee2ea-db16-4f99-9e06-43d8c932ad01 req-63672a65-cc54-4b6f-9482-c9d609ff4cdd service nova] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.862930] env[61906]: DEBUG nova.network.neutron [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.957913] env[61906]: DEBUG nova.network.neutron [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.961036] env[61906]: DEBUG nova.network.neutron [-] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.171058] env[61906]: DEBUG oslo_concurrency.lockutils [req-ebaee2ea-db16-4f99-9e06-43d8c932ad01 req-63672a65-cc54-4b6f-9482-c9d609ff4cdd service nova] Releasing lock "refresh_cache-2e8d82fd-5f70-408e-9d3b-74f3d81232d3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.204667] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69f86272-9c11-4a4f-b13b-7da258d61757 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.212459] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8df23298-838c-4001-8de0-546a27986a5e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.242450] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a995f466-7cf6-415b-858c-d1d0f15a3f2c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.249219] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba82936a-9812-4a77-a74b-428a6ae33cb5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.261576] env[61906]: DEBUG nova.compute.provider_tree [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 699.461112] env[61906]: DEBUG oslo_concurrency.lockutils [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Releasing lock "refresh_cache-32e2f482-9ceb-44bc-8933-115876374b1c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.461300] env[61906]: DEBUG nova.compute.manager [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 699.461486] env[61906]: DEBUG nova.compute.manager [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 699.461719] env[61906]: DEBUG nova.network.neutron [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 699.463496] env[61906]: INFO nova.compute.manager [-] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Took 1.02 seconds to deallocate network for instance. [ 699.478319] env[61906]: DEBUG nova.network.neutron [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 699.764259] env[61906]: DEBUG nova.scheduler.client.report [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 699.969766] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.981034] env[61906]: DEBUG nova.network.neutron [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.109755] env[61906]: DEBUG nova.compute.manager [req-d83f87c5-d398-4020-af90-85687364bdfa req-7fe1e684-0eaf-47a1-90f1-9c982a985a05 service nova] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Received event network-vif-deleted-99389367-21dd-45ae-8f2e-c1002dc479dd {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 700.156558] env[61906]: DEBUG nova.compute.manager [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 700.159731] env[61906]: DEBUG nova.virt.hardware [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 700.159731] env[61906]: DEBUG nova.virt.hardware [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 700.159731] env[61906]: DEBUG nova.virt.hardware [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 700.159731] env[61906]: DEBUG nova.virt.hardware [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 700.160147] env[61906]: DEBUG nova.virt.hardware [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 700.160147] env[61906]: DEBUG nova.virt.hardware [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 700.160147] env[61906]: DEBUG nova.virt.hardware [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 700.160147] env[61906]: DEBUG nova.virt.hardware [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 700.160147] env[61906]: DEBUG nova.virt.hardware [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 700.160293] env[61906]: DEBUG nova.virt.hardware [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 700.160293] env[61906]: DEBUG nova.virt.hardware [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 700.160293] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a10f4ce6-0ff7-4a5c-8448-30f1ce4c6162 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.168015] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9d3740-8d30-4fc3-9c4e-e9e7ade9fbc2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.182197] env[61906]: ERROR nova.compute.manager [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 99389367-21dd-45ae-8f2e-c1002dc479dd, please check neutron logs for more information. [ 700.182197] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Traceback (most recent call last): [ 700.182197] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 700.182197] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] yield resources [ 700.182197] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 700.182197] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] self.driver.spawn(context, instance, image_meta, [ 700.182197] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 700.182197] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 700.182197] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 700.182197] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] vm_ref = self.build_virtual_machine(instance, [ 700.182197] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 700.182755] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] vif_infos = vmwarevif.get_vif_info(self._session, [ 700.182755] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 700.182755] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] for vif in network_info: [ 700.182755] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 700.182755] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] return self._sync_wrapper(fn, *args, **kwargs) [ 700.182755] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 700.182755] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] self.wait() [ 700.182755] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 700.182755] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] self[:] = self._gt.wait() [ 700.182755] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 700.182755] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] return self._exit_event.wait() [ 700.182755] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 700.182755] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] current.throw(*self._exc) [ 700.183117] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 700.183117] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] result = function(*args, **kwargs) [ 700.183117] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 700.183117] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] return func(*args, **kwargs) [ 700.183117] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 700.183117] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] raise e [ 700.183117] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 700.183117] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] nwinfo = self.network_api.allocate_for_instance( [ 700.183117] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 700.183117] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] created_port_ids = self._update_ports_for_instance( [ 700.183117] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 700.183117] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] with excutils.save_and_reraise_exception(): [ 700.183117] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 700.183461] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] self.force_reraise() [ 700.183461] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 700.183461] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] raise self.value [ 700.183461] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 700.183461] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] updated_port = self._update_port( [ 700.183461] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 700.183461] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] _ensure_no_port_binding_failure(port) [ 700.183461] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 700.183461] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] raise exception.PortBindingFailed(port_id=port['id']) [ 700.183461] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] nova.exception.PortBindingFailed: Binding failed for port 99389367-21dd-45ae-8f2e-c1002dc479dd, please check neutron logs for more information. [ 700.183461] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] [ 700.183461] env[61906]: INFO nova.compute.manager [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Terminating instance [ 700.186357] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Acquiring lock "refresh_cache-2e8d82fd-5f70-408e-9d3b-74f3d81232d3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.186528] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Acquired lock "refresh_cache-2e8d82fd-5f70-408e-9d3b-74f3d81232d3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.186700] env[61906]: DEBUG nova.network.neutron [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 700.268822] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.939s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.269944] env[61906]: ERROR nova.compute.manager [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 75dff989-71c3-4b84-811d-eb77b29ac72d, please check neutron logs for more information. [ 700.269944] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Traceback (most recent call last): [ 700.269944] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 700.269944] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] self.driver.spawn(context, instance, image_meta, [ 700.269944] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 700.269944] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] self._vmops.spawn(context, instance, image_meta, injected_files, [ 700.269944] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 700.269944] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] vm_ref = self.build_virtual_machine(instance, [ 700.269944] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 700.269944] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] vif_infos = vmwarevif.get_vif_info(self._session, [ 700.269944] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 700.270321] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] for vif in network_info: [ 700.270321] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 700.270321] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] return self._sync_wrapper(fn, *args, **kwargs) [ 700.270321] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 700.270321] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] self.wait() [ 700.270321] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 700.270321] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] self[:] = self._gt.wait() [ 700.270321] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 700.270321] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] return self._exit_event.wait() [ 700.270321] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 700.270321] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] current.throw(*self._exc) [ 700.270321] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 700.270321] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] result = function(*args, **kwargs) [ 700.270698] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 700.270698] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] return func(*args, **kwargs) [ 700.270698] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 700.270698] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] raise e [ 700.270698] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 700.270698] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] nwinfo = self.network_api.allocate_for_instance( [ 700.270698] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 700.270698] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] created_port_ids = self._update_ports_for_instance( [ 700.270698] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 700.270698] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] with excutils.save_and_reraise_exception(): [ 700.270698] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 700.270698] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] self.force_reraise() [ 700.270698] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 700.271012] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] raise self.value [ 700.271012] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 700.271012] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] updated_port = self._update_port( [ 700.271012] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 700.271012] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] _ensure_no_port_binding_failure(port) [ 700.271012] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 700.271012] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] raise exception.PortBindingFailed(port_id=port['id']) [ 700.271012] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] nova.exception.PortBindingFailed: Binding failed for port 75dff989-71c3-4b84-811d-eb77b29ac72d, please check neutron logs for more information. [ 700.271012] env[61906]: ERROR nova.compute.manager [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] [ 700.271012] env[61906]: DEBUG nova.compute.utils [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Binding failed for port 75dff989-71c3-4b84-811d-eb77b29ac72d, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 700.271993] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.420s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 700.273802] env[61906]: INFO nova.compute.claims [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 700.276301] env[61906]: DEBUG nova.compute.manager [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Build of instance b6f46e45-742b-4801-86f0-fd8eea3611da was re-scheduled: Binding failed for port 75dff989-71c3-4b84-811d-eb77b29ac72d, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 700.276740] env[61906]: DEBUG nova.compute.manager [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 700.276924] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Acquiring lock "refresh_cache-b6f46e45-742b-4801-86f0-fd8eea3611da" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.277077] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Acquired lock "refresh_cache-b6f46e45-742b-4801-86f0-fd8eea3611da" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.277236] env[61906]: DEBUG nova.network.neutron [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 700.483659] env[61906]: INFO nova.compute.manager [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] [instance: 32e2f482-9ceb-44bc-8933-115876374b1c] Took 1.02 seconds to deallocate network for instance. [ 700.702357] env[61906]: DEBUG nova.network.neutron [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.786778] env[61906]: DEBUG nova.network.neutron [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.801341] env[61906]: DEBUG nova.network.neutron [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.876184] env[61906]: DEBUG nova.network.neutron [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.289623] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Releasing lock "refresh_cache-2e8d82fd-5f70-408e-9d3b-74f3d81232d3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.290171] env[61906]: DEBUG nova.compute.manager [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 701.290474] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3880e2e6-edd4-4b2a-8755-b07406d9c0fa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.299524] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d37d1d2-66d8-4a9a-82f2-2ff58f736c1b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.325579] env[61906]: WARNING nova.virt.vmwareapi.driver [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 2e8d82fd-5f70-408e-9d3b-74f3d81232d3 could not be found. [ 701.325863] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 701.328460] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d6a0eb63-f24f-47c7-b55b-640bcd3bab0f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.336896] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d21caa-4d36-4c5e-9739-c417f7922995 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.359413] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2e8d82fd-5f70-408e-9d3b-74f3d81232d3 could not be found. [ 701.359632] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 701.359815] env[61906]: INFO nova.compute.manager [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Took 0.07 seconds to destroy the instance on the hypervisor. [ 701.360063] env[61906]: DEBUG oslo.service.loopingcall [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 701.362349] env[61906]: DEBUG nova.compute.manager [-] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 701.362453] env[61906]: DEBUG nova.network.neutron [-] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 701.377809] env[61906]: DEBUG nova.network.neutron [-] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.379157] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Releasing lock "refresh_cache-b6f46e45-742b-4801-86f0-fd8eea3611da" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.379364] env[61906]: DEBUG nova.compute.manager [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 701.379536] env[61906]: DEBUG nova.compute.manager [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 701.379696] env[61906]: DEBUG nova.network.neutron [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 701.394060] env[61906]: DEBUG nova.network.neutron [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.520387] env[61906]: INFO nova.scheduler.client.report [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Deleted allocations for instance 32e2f482-9ceb-44bc-8933-115876374b1c [ 701.665325] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd75a24-720d-4796-a9f3-e7d2ff80084e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.673198] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d1363a5-f5fe-4844-8e0b-6b7943b55549 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.706302] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97226f6e-356c-4c57-bb9f-be45fd3fafeb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.713932] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb6e88e6-e492-4f70-8993-4a52983dbcc2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.727350] env[61906]: DEBUG nova.compute.provider_tree [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 701.881966] env[61906]: DEBUG nova.network.neutron [-] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.898715] env[61906]: DEBUG nova.network.neutron [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.030130] env[61906]: DEBUG oslo_concurrency.lockutils [None req-617df8dc-3127-4161-8091-0deba4ec7e02 tempest-ServersTestJSON-1433707305 tempest-ServersTestJSON-1433707305-project-member] Lock "32e2f482-9ceb-44bc-8933-115876374b1c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.904s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.231029] env[61906]: DEBUG nova.scheduler.client.report [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 702.385080] env[61906]: INFO nova.compute.manager [-] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Took 1.02 seconds to deallocate network for instance. [ 702.401880] env[61906]: INFO nova.compute.manager [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] [instance: b6f46e45-742b-4801-86f0-fd8eea3611da] Took 1.02 seconds to deallocate network for instance. [ 702.532924] env[61906]: DEBUG nova.compute.manager [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 702.736179] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.464s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.736730] env[61906]: DEBUG nova.compute.manager [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 702.740027] env[61906]: DEBUG oslo_concurrency.lockutils [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.764s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.947066] env[61906]: INFO nova.compute.manager [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Took 0.56 seconds to detach 1 volumes for instance. [ 702.948591] env[61906]: DEBUG nova.compute.claims [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 702.948755] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 703.058411] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 703.244203] env[61906]: DEBUG nova.compute.utils [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 703.249388] env[61906]: DEBUG nova.compute.manager [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 703.249388] env[61906]: DEBUG nova.network.neutron [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 703.295755] env[61906]: DEBUG nova.policy [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e89d45df95134c709b569d06744dea5e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1b33dea0794a48f78b9f519cb269a8c2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 703.450384] env[61906]: INFO nova.scheduler.client.report [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Deleted allocations for instance b6f46e45-742b-4801-86f0-fd8eea3611da [ 703.665995] env[61906]: DEBUG nova.network.neutron [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Successfully created port: 5e472a5c-dd42-4934-a6ad-553423b8f024 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 703.675588] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d13733b3-b19e-4e66-9cd8-6cfcdffbd4c2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.683527] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e3fe448-4e56-4c40-b07d-65a5737314b7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.719129] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59fa9555-e5aa-4974-bb43-fe1ceb17382f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.727434] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2acbc655-47b8-4a60-8d8a-1087e7baf638 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.741978] env[61906]: DEBUG nova.compute.provider_tree [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 703.749156] env[61906]: DEBUG nova.compute.manager [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 703.962640] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6016147d-d674-4b9f-86f5-96877eb0bf77 tempest-ServersTestFqdnHostnames-1036707341 tempest-ServersTestFqdnHostnames-1036707341-project-member] Lock "b6f46e45-742b-4801-86f0-fd8eea3611da" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.932s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.248509] env[61906]: DEBUG nova.scheduler.client.report [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 704.468155] env[61906]: DEBUG nova.compute.manager [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 704.752925] env[61906]: DEBUG oslo_concurrency.lockutils [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.013s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.753608] env[61906]: ERROR nova.compute.manager [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ead9d36e-fa05-4276-b0f1-2e9e005fe203, please check neutron logs for more information. [ 704.753608] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Traceback (most recent call last): [ 704.753608] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 704.753608] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] self.driver.spawn(context, instance, image_meta, [ 704.753608] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 704.753608] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 704.753608] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 704.753608] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] vm_ref = self.build_virtual_machine(instance, [ 704.753608] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 704.753608] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] vif_infos = vmwarevif.get_vif_info(self._session, [ 704.753608] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 704.753931] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] for vif in network_info: [ 704.753931] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 704.753931] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] return self._sync_wrapper(fn, *args, **kwargs) [ 704.753931] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 704.753931] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] self.wait() [ 704.753931] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 704.753931] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] self[:] = self._gt.wait() [ 704.753931] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 704.753931] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] return self._exit_event.wait() [ 704.753931] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 704.753931] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] result = hub.switch() [ 704.753931] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 704.753931] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] return self.greenlet.switch() [ 704.754258] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 704.754258] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] result = function(*args, **kwargs) [ 704.754258] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 704.754258] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] return func(*args, **kwargs) [ 704.754258] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 704.754258] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] raise e [ 704.754258] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 704.754258] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] nwinfo = self.network_api.allocate_for_instance( [ 704.754258] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 704.754258] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] created_port_ids = self._update_ports_for_instance( [ 704.754258] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 704.754258] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] with excutils.save_and_reraise_exception(): [ 704.754258] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.754637] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] self.force_reraise() [ 704.754637] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.754637] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] raise self.value [ 704.754637] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 704.754637] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] updated_port = self._update_port( [ 704.754637] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.754637] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] _ensure_no_port_binding_failure(port) [ 704.754637] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.754637] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] raise exception.PortBindingFailed(port_id=port['id']) [ 704.754637] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] nova.exception.PortBindingFailed: Binding failed for port ead9d36e-fa05-4276-b0f1-2e9e005fe203, please check neutron logs for more information. [ 704.754637] env[61906]: ERROR nova.compute.manager [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] [ 704.754933] env[61906]: DEBUG nova.compute.utils [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Binding failed for port ead9d36e-fa05-4276-b0f1-2e9e005fe203, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 704.758580] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.020s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.759157] env[61906]: DEBUG nova.compute.manager [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Build of instance 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb was re-scheduled: Binding failed for port ead9d36e-fa05-4276-b0f1-2e9e005fe203, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 704.760950] env[61906]: DEBUG nova.compute.manager [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 704.761219] env[61906]: DEBUG oslo_concurrency.lockutils [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Acquiring lock "refresh_cache-0a1b97fe-42fb-47bd-bba2-aeb75c0710fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.761370] env[61906]: DEBUG oslo_concurrency.lockutils [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Acquired lock "refresh_cache-0a1b97fe-42fb-47bd-bba2-aeb75c0710fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.761528] env[61906]: DEBUG nova.network.neutron [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 704.765229] env[61906]: DEBUG nova.compute.manager [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 704.804798] env[61906]: DEBUG nova.virt.hardware [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 704.804798] env[61906]: DEBUG nova.virt.hardware [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 704.805157] env[61906]: DEBUG nova.virt.hardware [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 704.805157] env[61906]: DEBUG nova.virt.hardware [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 704.806041] env[61906]: DEBUG nova.virt.hardware [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 704.806231] env[61906]: DEBUG nova.virt.hardware [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 704.806434] env[61906]: DEBUG nova.virt.hardware [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 704.806596] env[61906]: DEBUG nova.virt.hardware [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 704.806764] env[61906]: DEBUG nova.virt.hardware [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 704.806965] env[61906]: DEBUG nova.virt.hardware [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 704.807100] env[61906]: DEBUG nova.virt.hardware [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 704.808474] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86294041-3f7a-4aee-8209-98d9f1e9faa6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.821014] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64c99f20-a1f0-48bf-a115-b7f3b0df9ce3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.959617] env[61906]: DEBUG nova.compute.manager [req-2eabdb35-d5c7-47d2-bcaa-101e99746101 req-3d0a0a1f-1939-4ee0-a623-33f6db95b9a1 service nova] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Received event network-changed-5e472a5c-dd42-4934-a6ad-553423b8f024 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 704.959617] env[61906]: DEBUG nova.compute.manager [req-2eabdb35-d5c7-47d2-bcaa-101e99746101 req-3d0a0a1f-1939-4ee0-a623-33f6db95b9a1 service nova] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Refreshing instance network info cache due to event network-changed-5e472a5c-dd42-4934-a6ad-553423b8f024. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 704.959617] env[61906]: DEBUG oslo_concurrency.lockutils [req-2eabdb35-d5c7-47d2-bcaa-101e99746101 req-3d0a0a1f-1939-4ee0-a623-33f6db95b9a1 service nova] Acquiring lock "refresh_cache-fd30cb23-f30f-4e20-ae67-36956b78e6ac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.959617] env[61906]: DEBUG oslo_concurrency.lockutils [req-2eabdb35-d5c7-47d2-bcaa-101e99746101 req-3d0a0a1f-1939-4ee0-a623-33f6db95b9a1 service nova] Acquired lock "refresh_cache-fd30cb23-f30f-4e20-ae67-36956b78e6ac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.959617] env[61906]: DEBUG nova.network.neutron [req-2eabdb35-d5c7-47d2-bcaa-101e99746101 req-3d0a0a1f-1939-4ee0-a623-33f6db95b9a1 service nova] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Refreshing network info cache for port 5e472a5c-dd42-4934-a6ad-553423b8f024 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 704.991948] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.118816] env[61906]: ERROR nova.compute.manager [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5e472a5c-dd42-4934-a6ad-553423b8f024, please check neutron logs for more information. [ 705.118816] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 705.118816] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.118816] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 705.118816] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 705.118816] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 705.118816] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 705.118816] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 705.118816] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.118816] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 705.118816] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.118816] env[61906]: ERROR nova.compute.manager raise self.value [ 705.118816] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 705.118816] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 705.118816] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.118816] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 705.119290] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.119290] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 705.119290] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5e472a5c-dd42-4934-a6ad-553423b8f024, please check neutron logs for more information. [ 705.119290] env[61906]: ERROR nova.compute.manager [ 705.119290] env[61906]: Traceback (most recent call last): [ 705.119290] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 705.119290] env[61906]: listener.cb(fileno) [ 705.119290] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.119290] env[61906]: result = function(*args, **kwargs) [ 705.119290] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 705.119290] env[61906]: return func(*args, **kwargs) [ 705.119290] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 705.119290] env[61906]: raise e [ 705.119290] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.119290] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 705.119290] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 705.119290] env[61906]: created_port_ids = self._update_ports_for_instance( [ 705.119290] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 705.119290] env[61906]: with excutils.save_and_reraise_exception(): [ 705.119290] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.119290] env[61906]: self.force_reraise() [ 705.119290] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.119290] env[61906]: raise self.value [ 705.119290] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 705.119290] env[61906]: updated_port = self._update_port( [ 705.119290] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.119290] env[61906]: _ensure_no_port_binding_failure(port) [ 705.119290] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.119290] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 705.120102] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 5e472a5c-dd42-4934-a6ad-553423b8f024, please check neutron logs for more information. [ 705.120102] env[61906]: Removing descriptor: 16 [ 705.120102] env[61906]: ERROR nova.compute.manager [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5e472a5c-dd42-4934-a6ad-553423b8f024, please check neutron logs for more information. [ 705.120102] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Traceback (most recent call last): [ 705.120102] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 705.120102] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] yield resources [ 705.120102] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 705.120102] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] self.driver.spawn(context, instance, image_meta, [ 705.120102] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 705.120102] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] self._vmops.spawn(context, instance, image_meta, injected_files, [ 705.120102] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 705.120102] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] vm_ref = self.build_virtual_machine(instance, [ 705.120508] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 705.120508] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] vif_infos = vmwarevif.get_vif_info(self._session, [ 705.120508] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 705.120508] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] for vif in network_info: [ 705.120508] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 705.120508] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] return self._sync_wrapper(fn, *args, **kwargs) [ 705.120508] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 705.120508] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] self.wait() [ 705.120508] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 705.120508] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] self[:] = self._gt.wait() [ 705.120508] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 705.120508] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] return self._exit_event.wait() [ 705.120508] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 705.120882] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] result = hub.switch() [ 705.120882] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 705.120882] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] return self.greenlet.switch() [ 705.120882] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.120882] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] result = function(*args, **kwargs) [ 705.120882] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 705.120882] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] return func(*args, **kwargs) [ 705.120882] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 705.120882] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] raise e [ 705.120882] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.120882] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] nwinfo = self.network_api.allocate_for_instance( [ 705.120882] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 705.120882] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] created_port_ids = self._update_ports_for_instance( [ 705.121282] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 705.121282] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] with excutils.save_and_reraise_exception(): [ 705.121282] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.121282] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] self.force_reraise() [ 705.121282] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.121282] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] raise self.value [ 705.121282] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 705.121282] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] updated_port = self._update_port( [ 705.121282] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.121282] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] _ensure_no_port_binding_failure(port) [ 705.121282] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.121282] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] raise exception.PortBindingFailed(port_id=port['id']) [ 705.121605] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] nova.exception.PortBindingFailed: Binding failed for port 5e472a5c-dd42-4934-a6ad-553423b8f024, please check neutron logs for more information. [ 705.121605] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] [ 705.121605] env[61906]: INFO nova.compute.manager [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Terminating instance [ 705.125371] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "refresh_cache-fd30cb23-f30f-4e20-ae67-36956b78e6ac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.297955] env[61906]: DEBUG nova.network.neutron [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.485462] env[61906]: DEBUG nova.network.neutron [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.489437] env[61906]: DEBUG nova.network.neutron [req-2eabdb35-d5c7-47d2-bcaa-101e99746101 req-3d0a0a1f-1939-4ee0-a623-33f6db95b9a1 service nova] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.641952] env[61906]: DEBUG nova.network.neutron [req-2eabdb35-d5c7-47d2-bcaa-101e99746101 req-3d0a0a1f-1939-4ee0-a623-33f6db95b9a1 service nova] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.708451] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-980f16a5-3c5d-47a2-ad88-6ac3798b9b3b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.716581] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42716748-512c-4571-81dc-68d7fe2ddca0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.750413] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d9cd1f-9e10-44a9-a9c7-ca8ccfcfb43e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.758184] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0082ca5e-84b4-4807-8a8e-0a12a51b3682 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.771168] env[61906]: DEBUG nova.compute.provider_tree [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.991793] env[61906]: DEBUG oslo_concurrency.lockutils [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Releasing lock "refresh_cache-0a1b97fe-42fb-47bd-bba2-aeb75c0710fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.991793] env[61906]: DEBUG nova.compute.manager [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 705.991793] env[61906]: DEBUG nova.compute.manager [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 705.991793] env[61906]: DEBUG nova.network.neutron [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 706.010512] env[61906]: DEBUG nova.network.neutron [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.146102] env[61906]: DEBUG oslo_concurrency.lockutils [req-2eabdb35-d5c7-47d2-bcaa-101e99746101 req-3d0a0a1f-1939-4ee0-a623-33f6db95b9a1 service nova] Releasing lock "refresh_cache-fd30cb23-f30f-4e20-ae67-36956b78e6ac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.146102] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquired lock "refresh_cache-fd30cb23-f30f-4e20-ae67-36956b78e6ac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.146216] env[61906]: DEBUG nova.network.neutron [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 706.275059] env[61906]: DEBUG nova.scheduler.client.report [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 706.520376] env[61906]: DEBUG nova.network.neutron [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.673136] env[61906]: DEBUG nova.network.neutron [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.784194] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.028s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.784847] env[61906]: ERROR nova.compute.manager [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6479ad8a-7a5a-4806-a541-37a118722244, please check neutron logs for more information. [ 706.784847] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Traceback (most recent call last): [ 706.784847] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 706.784847] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] self.driver.spawn(context, instance, image_meta, [ 706.784847] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 706.784847] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] self._vmops.spawn(context, instance, image_meta, injected_files, [ 706.784847] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 706.784847] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] vm_ref = self.build_virtual_machine(instance, [ 706.784847] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 706.784847] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] vif_infos = vmwarevif.get_vif_info(self._session, [ 706.784847] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 706.785192] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] for vif in network_info: [ 706.785192] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 706.785192] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] return self._sync_wrapper(fn, *args, **kwargs) [ 706.785192] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 706.785192] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] self.wait() [ 706.785192] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 706.785192] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] self[:] = self._gt.wait() [ 706.785192] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 706.785192] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] return self._exit_event.wait() [ 706.785192] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 706.785192] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] result = hub.switch() [ 706.785192] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 706.785192] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] return self.greenlet.switch() [ 706.785621] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.785621] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] result = function(*args, **kwargs) [ 706.785621] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 706.785621] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] return func(*args, **kwargs) [ 706.785621] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 706.785621] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] raise e [ 706.785621] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.785621] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] nwinfo = self.network_api.allocate_for_instance( [ 706.785621] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 706.785621] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] created_port_ids = self._update_ports_for_instance( [ 706.785621] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 706.785621] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] with excutils.save_and_reraise_exception(): [ 706.785621] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.785993] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] self.force_reraise() [ 706.785993] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.785993] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] raise self.value [ 706.785993] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 706.785993] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] updated_port = self._update_port( [ 706.785993] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.785993] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] _ensure_no_port_binding_failure(port) [ 706.785993] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.785993] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] raise exception.PortBindingFailed(port_id=port['id']) [ 706.785993] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] nova.exception.PortBindingFailed: Binding failed for port 6479ad8a-7a5a-4806-a541-37a118722244, please check neutron logs for more information. [ 706.785993] env[61906]: ERROR nova.compute.manager [instance: 59887ef7-b2fc-486c-a368-cebde3313113] [ 706.786321] env[61906]: DEBUG nova.compute.utils [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Binding failed for port 6479ad8a-7a5a-4806-a541-37a118722244, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 706.786731] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.753s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.794021] env[61906]: DEBUG nova.compute.manager [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Build of instance 59887ef7-b2fc-486c-a368-cebde3313113 was re-scheduled: Binding failed for port 6479ad8a-7a5a-4806-a541-37a118722244, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 706.794021] env[61906]: DEBUG nova.compute.manager [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 706.794021] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Acquiring lock "refresh_cache-59887ef7-b2fc-486c-a368-cebde3313113" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.794021] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Acquired lock "refresh_cache-59887ef7-b2fc-486c-a368-cebde3313113" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.794269] env[61906]: DEBUG nova.network.neutron [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 706.840391] env[61906]: DEBUG nova.network.neutron [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.989762] env[61906]: DEBUG nova.compute.manager [req-6fed2da7-8dc2-4102-b0b6-369820bf0742 req-937b9c01-5d42-4edf-a056-5ea3a6352fe9 service nova] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Received event network-vif-deleted-5e472a5c-dd42-4934-a6ad-553423b8f024 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 707.022563] env[61906]: INFO nova.compute.manager [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb] Took 1.03 seconds to deallocate network for instance. [ 707.319686] env[61906]: DEBUG nova.network.neutron [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.343322] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Releasing lock "refresh_cache-fd30cb23-f30f-4e20-ae67-36956b78e6ac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.344615] env[61906]: DEBUG nova.compute.manager [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 707.344615] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 707.344615] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d44e02c5-2f47-4ea5-bf36-1e5e2ca6f9ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.355790] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbac46c4-68f9-4d2a-bb54-37d947b0570a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.383988] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fd30cb23-f30f-4e20-ae67-36956b78e6ac could not be found. [ 707.384278] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 707.384497] env[61906]: INFO nova.compute.manager [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Took 0.04 seconds to destroy the instance on the hypervisor. [ 707.384775] env[61906]: DEBUG oslo.service.loopingcall [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 707.387267] env[61906]: DEBUG nova.compute.manager [-] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 707.387507] env[61906]: DEBUG nova.network.neutron [-] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 707.414451] env[61906]: DEBUG nova.network.neutron [-] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.469613] env[61906]: DEBUG nova.network.neutron [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.784199] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19e358ae-c404-4cd2-b3fe-f3993facb520 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.792674] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e49ce9c-c7ef-424a-ab64-f9ebcebfb854 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.827710] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2844bf46-1a41-406a-a7db-09b18f382b3a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.841392] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c233768-47ed-43a1-a44f-a4332aff3392 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.856946] env[61906]: DEBUG nova.compute.provider_tree [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 707.918611] env[61906]: DEBUG nova.network.neutron [-] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.974716] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Releasing lock "refresh_cache-59887ef7-b2fc-486c-a368-cebde3313113" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.975008] env[61906]: DEBUG nova.compute.manager [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 707.975214] env[61906]: DEBUG nova.compute.manager [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 707.975395] env[61906]: DEBUG nova.network.neutron [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 707.995375] env[61906]: DEBUG nova.network.neutron [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.056485] env[61906]: INFO nova.scheduler.client.report [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Deleted allocations for instance 0a1b97fe-42fb-47bd-bba2-aeb75c0710fb [ 708.360489] env[61906]: DEBUG nova.scheduler.client.report [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 708.421540] env[61906]: INFO nova.compute.manager [-] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Took 1.03 seconds to deallocate network for instance. [ 708.426360] env[61906]: DEBUG nova.compute.claims [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 708.426360] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.497718] env[61906]: DEBUG nova.network.neutron [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.565911] env[61906]: DEBUG oslo_concurrency.lockutils [None req-118fbc1a-5f14-49d4-af7b-a405ddc7bb91 tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Lock "0a1b97fe-42fb-47bd-bba2-aeb75c0710fb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.755s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.869067] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.082s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.869808] env[61906]: ERROR nova.compute.manager [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f3cfb5a1-a938-47ba-8043-bb3f72ab73fd, please check neutron logs for more information. [ 708.869808] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Traceback (most recent call last): [ 708.869808] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 708.869808] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] self.driver.spawn(context, instance, image_meta, [ 708.869808] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 708.869808] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] self._vmops.spawn(context, instance, image_meta, injected_files, [ 708.869808] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 708.869808] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] vm_ref = self.build_virtual_machine(instance, [ 708.869808] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 708.869808] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] vif_infos = vmwarevif.get_vif_info(self._session, [ 708.869808] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 708.870122] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] for vif in network_info: [ 708.870122] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 708.870122] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] return self._sync_wrapper(fn, *args, **kwargs) [ 708.870122] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 708.870122] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] self.wait() [ 708.870122] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 708.870122] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] self[:] = self._gt.wait() [ 708.870122] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 708.870122] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] return self._exit_event.wait() [ 708.870122] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 708.870122] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] current.throw(*self._exc) [ 708.870122] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 708.870122] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] result = function(*args, **kwargs) [ 708.870435] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 708.870435] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] return func(*args, **kwargs) [ 708.870435] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 708.870435] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] raise e [ 708.870435] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 708.870435] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] nwinfo = self.network_api.allocate_for_instance( [ 708.870435] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 708.870435] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] created_port_ids = self._update_ports_for_instance( [ 708.870435] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 708.870435] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] with excutils.save_and_reraise_exception(): [ 708.870435] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.870435] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] self.force_reraise() [ 708.870435] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.870835] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] raise self.value [ 708.870835] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 708.870835] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] updated_port = self._update_port( [ 708.870835] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.870835] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] _ensure_no_port_binding_failure(port) [ 708.870835] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.870835] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] raise exception.PortBindingFailed(port_id=port['id']) [ 708.870835] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] nova.exception.PortBindingFailed: Binding failed for port f3cfb5a1-a938-47ba-8043-bb3f72ab73fd, please check neutron logs for more information. [ 708.870835] env[61906]: ERROR nova.compute.manager [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] [ 708.870835] env[61906]: DEBUG nova.compute.utils [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Binding failed for port f3cfb5a1-a938-47ba-8043-bb3f72ab73fd, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 708.871964] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.211s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.874243] env[61906]: INFO nova.compute.claims [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 708.877269] env[61906]: DEBUG nova.compute.manager [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Build of instance ff6fa10c-c1cc-4e85-af43-c52f06e35974 was re-scheduled: Binding failed for port f3cfb5a1-a938-47ba-8043-bb3f72ab73fd, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 708.877710] env[61906]: DEBUG nova.compute.manager [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 708.877935] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Acquiring lock "refresh_cache-ff6fa10c-c1cc-4e85-af43-c52f06e35974" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.878098] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Acquired lock "refresh_cache-ff6fa10c-c1cc-4e85-af43-c52f06e35974" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.878258] env[61906]: DEBUG nova.network.neutron [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 709.000444] env[61906]: INFO nova.compute.manager [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] [instance: 59887ef7-b2fc-486c-a368-cebde3313113] Took 1.02 seconds to deallocate network for instance. [ 709.069094] env[61906]: DEBUG nova.compute.manager [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 709.412671] env[61906]: DEBUG nova.network.neutron [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 709.532296] env[61906]: DEBUG nova.network.neutron [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.595700] env[61906]: DEBUG oslo_concurrency.lockutils [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.034759] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Releasing lock "refresh_cache-ff6fa10c-c1cc-4e85-af43-c52f06e35974" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.035504] env[61906]: DEBUG nova.compute.manager [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 710.036984] env[61906]: DEBUG nova.compute.manager [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 710.036984] env[61906]: DEBUG nova.network.neutron [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 710.039491] env[61906]: INFO nova.scheduler.client.report [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Deleted allocations for instance 59887ef7-b2fc-486c-a368-cebde3313113 [ 710.077977] env[61906]: DEBUG nova.network.neutron [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.425369] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbcc929f-927e-4f5c-95a7-323dbc5d22fc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.440519] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e1597d2-dbb1-4019-b0ef-8faaa45da480 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.483323] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c659d77d-2be6-44a7-b9e7-a0a6ccc06353 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.492422] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1dc0b98-1856-49a9-9c85-8f8c7c3405f2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.507411] env[61906]: DEBUG nova.compute.provider_tree [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 710.552406] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74973d38-0779-42d1-8810-f29ff8d5a516 tempest-InstanceActionsTestJSON-1756675091 tempest-InstanceActionsTestJSON-1756675091-project-member] Lock "59887ef7-b2fc-486c-a368-cebde3313113" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.396s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.583778] env[61906]: DEBUG nova.network.neutron [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.013109] env[61906]: DEBUG nova.scheduler.client.report [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 711.020945] env[61906]: DEBUG oslo_concurrency.lockutils [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Acquiring lock "52a387d2-4090-4eba-9a77-98c37f152b75" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.021358] env[61906]: DEBUG oslo_concurrency.lockutils [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Lock "52a387d2-4090-4eba-9a77-98c37f152b75" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.058931] env[61906]: DEBUG nova.compute.manager [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 711.089737] env[61906]: INFO nova.compute.manager [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] [instance: ff6fa10c-c1cc-4e85-af43-c52f06e35974] Took 1.05 seconds to deallocate network for instance. [ 711.521535] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.649s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.522055] env[61906]: DEBUG nova.compute.manager [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 711.525124] env[61906]: DEBUG oslo_concurrency.lockutils [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.937s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.530428] env[61906]: INFO nova.compute.claims [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 711.591430] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.035087] env[61906]: DEBUG nova.compute.utils [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 712.038428] env[61906]: DEBUG nova.compute.manager [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 712.038649] env[61906]: DEBUG nova.network.neutron [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 712.120542] env[61906]: INFO nova.scheduler.client.report [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Deleted allocations for instance ff6fa10c-c1cc-4e85-af43-c52f06e35974 [ 712.127536] env[61906]: DEBUG nova.policy [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e785986dfcfc4f4f91bc786138fcfcc9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f2265e8dba0647238f355f1e98008150', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 712.539773] env[61906]: DEBUG nova.compute.manager [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 712.623502] env[61906]: DEBUG nova.network.neutron [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Successfully created port: 774f613b-ada5-476b-95e0-cbb8bb0f4da4 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 712.630951] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6eeb0c7b-e0ce-4c34-9d80-4522b106812f tempest-ServersAdminTestJSON-1763435921 tempest-ServersAdminTestJSON-1763435921-project-member] Lock "ff6fa10c-c1cc-4e85-af43-c52f06e35974" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 129.359s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.974020] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7449d48-d82c-4851-8144-2469a04c8f9b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.980386] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d55c0b1-5c65-48c2-88d3-40982389080d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.020050] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-442975b6-bcbf-4293-8982-f51cabbec25e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.027939] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa052dd0-42bd-48f1-9f76-b8ebdcbacdfe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.043586] env[61906]: DEBUG nova.compute.provider_tree [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.137782] env[61906]: DEBUG nova.compute.manager [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 713.547518] env[61906]: DEBUG nova.scheduler.client.report [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 713.555722] env[61906]: DEBUG nova.compute.manager [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 713.578166] env[61906]: DEBUG nova.compute.manager [req-df9c2d13-d546-4169-adf0-fd195cdfa422 req-665f552a-9757-4e48-8073-699dfdd540cd service nova] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Received event network-changed-774f613b-ada5-476b-95e0-cbb8bb0f4da4 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 713.578166] env[61906]: DEBUG nova.compute.manager [req-df9c2d13-d546-4169-adf0-fd195cdfa422 req-665f552a-9757-4e48-8073-699dfdd540cd service nova] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Refreshing instance network info cache due to event network-changed-774f613b-ada5-476b-95e0-cbb8bb0f4da4. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 713.578166] env[61906]: DEBUG oslo_concurrency.lockutils [req-df9c2d13-d546-4169-adf0-fd195cdfa422 req-665f552a-9757-4e48-8073-699dfdd540cd service nova] Acquiring lock "refresh_cache-fe14efda-3b92-4045-8eef-0e5e92a9c538" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.578166] env[61906]: DEBUG oslo_concurrency.lockutils [req-df9c2d13-d546-4169-adf0-fd195cdfa422 req-665f552a-9757-4e48-8073-699dfdd540cd service nova] Acquired lock "refresh_cache-fe14efda-3b92-4045-8eef-0e5e92a9c538" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.578166] env[61906]: DEBUG nova.network.neutron [req-df9c2d13-d546-4169-adf0-fd195cdfa422 req-665f552a-9757-4e48-8073-699dfdd540cd service nova] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Refreshing network info cache for port 774f613b-ada5-476b-95e0-cbb8bb0f4da4 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 713.590087] env[61906]: DEBUG nova.virt.hardware [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 713.590344] env[61906]: DEBUG nova.virt.hardware [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 713.590526] env[61906]: DEBUG nova.virt.hardware [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 713.590746] env[61906]: DEBUG nova.virt.hardware [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 713.590886] env[61906]: DEBUG nova.virt.hardware [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 713.591100] env[61906]: DEBUG nova.virt.hardware [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 713.591429] env[61906]: DEBUG nova.virt.hardware [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 713.591559] env[61906]: DEBUG nova.virt.hardware [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 713.591796] env[61906]: DEBUG nova.virt.hardware [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 713.591961] env[61906]: DEBUG nova.virt.hardware [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 713.592160] env[61906]: DEBUG nova.virt.hardware [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 713.594418] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfa1366e-d674-4ab8-9414-5c5e9c764a7b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.604854] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21ec8d2d-a798-4eb8-a929-386f26ba9755 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.664903] env[61906]: DEBUG oslo_concurrency.lockutils [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.769263] env[61906]: ERROR nova.compute.manager [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 774f613b-ada5-476b-95e0-cbb8bb0f4da4, please check neutron logs for more information. [ 713.769263] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 713.769263] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 713.769263] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 713.769263] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.769263] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 713.769263] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.769263] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 713.769263] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.769263] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 713.769263] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.769263] env[61906]: ERROR nova.compute.manager raise self.value [ 713.769263] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.769263] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 713.769263] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.769263] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 713.769733] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.769733] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 713.769733] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 774f613b-ada5-476b-95e0-cbb8bb0f4da4, please check neutron logs for more information. [ 713.769733] env[61906]: ERROR nova.compute.manager [ 713.769733] env[61906]: Traceback (most recent call last): [ 713.769733] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 713.769733] env[61906]: listener.cb(fileno) [ 713.769733] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 713.769733] env[61906]: result = function(*args, **kwargs) [ 713.769733] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 713.769733] env[61906]: return func(*args, **kwargs) [ 713.769733] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 713.769733] env[61906]: raise e [ 713.769733] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 713.769733] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 713.769733] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.769733] env[61906]: created_port_ids = self._update_ports_for_instance( [ 713.769733] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.769733] env[61906]: with excutils.save_and_reraise_exception(): [ 713.769733] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.769733] env[61906]: self.force_reraise() [ 713.769733] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.769733] env[61906]: raise self.value [ 713.769733] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.769733] env[61906]: updated_port = self._update_port( [ 713.769733] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.769733] env[61906]: _ensure_no_port_binding_failure(port) [ 713.769733] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.769733] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 713.770445] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 774f613b-ada5-476b-95e0-cbb8bb0f4da4, please check neutron logs for more information. [ 713.770445] env[61906]: Removing descriptor: 18 [ 713.770445] env[61906]: ERROR nova.compute.manager [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 774f613b-ada5-476b-95e0-cbb8bb0f4da4, please check neutron logs for more information. [ 713.770445] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Traceback (most recent call last): [ 713.770445] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 713.770445] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] yield resources [ 713.770445] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 713.770445] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] self.driver.spawn(context, instance, image_meta, [ 713.770445] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 713.770445] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] self._vmops.spawn(context, instance, image_meta, injected_files, [ 713.770445] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 713.770445] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] vm_ref = self.build_virtual_machine(instance, [ 713.770785] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 713.770785] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] vif_infos = vmwarevif.get_vif_info(self._session, [ 713.770785] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 713.770785] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] for vif in network_info: [ 713.770785] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 713.770785] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] return self._sync_wrapper(fn, *args, **kwargs) [ 713.770785] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 713.770785] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] self.wait() [ 713.770785] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 713.770785] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] self[:] = self._gt.wait() [ 713.770785] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 713.770785] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] return self._exit_event.wait() [ 713.770785] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 713.771152] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] result = hub.switch() [ 713.771152] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 713.771152] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] return self.greenlet.switch() [ 713.771152] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 713.771152] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] result = function(*args, **kwargs) [ 713.771152] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 713.771152] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] return func(*args, **kwargs) [ 713.771152] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 713.771152] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] raise e [ 713.771152] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 713.771152] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] nwinfo = self.network_api.allocate_for_instance( [ 713.771152] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.771152] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] created_port_ids = self._update_ports_for_instance( [ 713.771509] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.771509] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] with excutils.save_and_reraise_exception(): [ 713.771509] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.771509] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] self.force_reraise() [ 713.771509] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.771509] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] raise self.value [ 713.771509] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.771509] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] updated_port = self._update_port( [ 713.771509] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.771509] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] _ensure_no_port_binding_failure(port) [ 713.771509] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.771509] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] raise exception.PortBindingFailed(port_id=port['id']) [ 713.771911] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] nova.exception.PortBindingFailed: Binding failed for port 774f613b-ada5-476b-95e0-cbb8bb0f4da4, please check neutron logs for more information. [ 713.771911] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] [ 713.771911] env[61906]: INFO nova.compute.manager [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Terminating instance [ 713.773116] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Acquiring lock "refresh_cache-fe14efda-3b92-4045-8eef-0e5e92a9c538" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.059576] env[61906]: DEBUG oslo_concurrency.lockutils [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.534s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.060179] env[61906]: DEBUG nova.compute.manager [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 714.063675] env[61906]: DEBUG oslo_concurrency.lockutils [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.833s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.065057] env[61906]: INFO nova.compute.claims [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 714.104545] env[61906]: DEBUG nova.network.neutron [req-df9c2d13-d546-4169-adf0-fd195cdfa422 req-665f552a-9757-4e48-8073-699dfdd540cd service nova] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.199630] env[61906]: DEBUG nova.network.neutron [req-df9c2d13-d546-4169-adf0-fd195cdfa422 req-665f552a-9757-4e48-8073-699dfdd540cd service nova] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.572403] env[61906]: DEBUG nova.compute.utils [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 714.573831] env[61906]: DEBUG nova.compute.manager [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 714.573978] env[61906]: DEBUG nova.network.neutron [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 714.621034] env[61906]: DEBUG nova.policy [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '45881b6e365f4ae8a0516916d4a89529', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ff2b69fe6514278a3c293c26432fb8b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 714.702159] env[61906]: DEBUG oslo_concurrency.lockutils [req-df9c2d13-d546-4169-adf0-fd195cdfa422 req-665f552a-9757-4e48-8073-699dfdd540cd service nova] Releasing lock "refresh_cache-fe14efda-3b92-4045-8eef-0e5e92a9c538" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.702583] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Acquired lock "refresh_cache-fe14efda-3b92-4045-8eef-0e5e92a9c538" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.702762] env[61906]: DEBUG nova.network.neutron [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 714.954104] env[61906]: DEBUG nova.network.neutron [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Successfully created port: ebd40d4d-5c50-4ebd-81a4-44ec4d9bfb2e {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 715.078534] env[61906]: DEBUG nova.compute.manager [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 715.233105] env[61906]: DEBUG nova.network.neutron [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.423133] env[61906]: DEBUG nova.network.neutron [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.503890] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf7eba5a-aed0-4de6-b351-0b223f007f93 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.512200] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44923039-816f-4a40-88e8-5a06605e2d9b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.541580] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03f516a8-e48d-4b66-baba-f29e49cdea88 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.549110] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f9500f6-00ab-481d-a0ec-826904079d52 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.562538] env[61906]: DEBUG nova.compute.provider_tree [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.818538] env[61906]: DEBUG nova.compute.manager [req-4476d23c-c896-47f3-904e-8d911c949fec req-cb9f9c2d-806e-42c1-921c-a1e925d59020 service nova] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Received event network-vif-deleted-774f613b-ada5-476b-95e0-cbb8bb0f4da4 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 715.926747] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Releasing lock "refresh_cache-fe14efda-3b92-4045-8eef-0e5e92a9c538" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.926747] env[61906]: DEBUG nova.compute.manager [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 715.926747] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 715.926971] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b2286f6e-7c31-4559-a071-d434162f3578 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.935943] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4341e55a-aea9-4eb3-ae92-5e600503acda {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.958015] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fe14efda-3b92-4045-8eef-0e5e92a9c538 could not be found. [ 715.958793] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 715.958793] env[61906]: INFO nova.compute.manager [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Took 0.03 seconds to destroy the instance on the hypervisor. [ 715.958793] env[61906]: DEBUG oslo.service.loopingcall [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 715.959192] env[61906]: DEBUG nova.compute.manager [-] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 715.959192] env[61906]: DEBUG nova.network.neutron [-] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 715.984123] env[61906]: DEBUG nova.network.neutron [-] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 716.067134] env[61906]: DEBUG nova.scheduler.client.report [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 716.088914] env[61906]: DEBUG nova.compute.manager [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 716.116140] env[61906]: DEBUG nova.virt.hardware [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 716.116140] env[61906]: DEBUG nova.virt.hardware [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 716.116366] env[61906]: DEBUG nova.virt.hardware [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 716.116917] env[61906]: DEBUG nova.virt.hardware [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 716.116917] env[61906]: DEBUG nova.virt.hardware [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 716.116917] env[61906]: DEBUG nova.virt.hardware [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 716.116917] env[61906]: DEBUG nova.virt.hardware [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 716.117115] env[61906]: DEBUG nova.virt.hardware [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 716.117234] env[61906]: DEBUG nova.virt.hardware [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 716.117395] env[61906]: DEBUG nova.virt.hardware [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 716.117570] env[61906]: DEBUG nova.virt.hardware [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 716.118437] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fab9a78-cda0-4755-b112-7f0c6c414fc1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.126800] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-064c7a34-5416-4488-b965-58ba0816b844 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.195146] env[61906]: ERROR nova.compute.manager [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ebd40d4d-5c50-4ebd-81a4-44ec4d9bfb2e, please check neutron logs for more information. [ 716.195146] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 716.195146] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.195146] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 716.195146] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 716.195146] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 716.195146] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 716.195146] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 716.195146] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.195146] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 716.195146] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.195146] env[61906]: ERROR nova.compute.manager raise self.value [ 716.195146] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 716.195146] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 716.195146] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.195146] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 716.195734] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.195734] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 716.195734] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ebd40d4d-5c50-4ebd-81a4-44ec4d9bfb2e, please check neutron logs for more information. [ 716.195734] env[61906]: ERROR nova.compute.manager [ 716.195734] env[61906]: Traceback (most recent call last): [ 716.195734] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 716.195734] env[61906]: listener.cb(fileno) [ 716.195734] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.195734] env[61906]: result = function(*args, **kwargs) [ 716.195734] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 716.195734] env[61906]: return func(*args, **kwargs) [ 716.195734] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 716.195734] env[61906]: raise e [ 716.195734] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.195734] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 716.195734] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 716.195734] env[61906]: created_port_ids = self._update_ports_for_instance( [ 716.195734] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 716.195734] env[61906]: with excutils.save_and_reraise_exception(): [ 716.195734] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.195734] env[61906]: self.force_reraise() [ 716.195734] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.195734] env[61906]: raise self.value [ 716.195734] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 716.195734] env[61906]: updated_port = self._update_port( [ 716.195734] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.195734] env[61906]: _ensure_no_port_binding_failure(port) [ 716.195734] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.195734] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 716.196545] env[61906]: nova.exception.PortBindingFailed: Binding failed for port ebd40d4d-5c50-4ebd-81a4-44ec4d9bfb2e, please check neutron logs for more information. [ 716.196545] env[61906]: Removing descriptor: 18 [ 716.196545] env[61906]: ERROR nova.compute.manager [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ebd40d4d-5c50-4ebd-81a4-44ec4d9bfb2e, please check neutron logs for more information. [ 716.196545] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Traceback (most recent call last): [ 716.196545] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 716.196545] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] yield resources [ 716.196545] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 716.196545] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] self.driver.spawn(context, instance, image_meta, [ 716.196545] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 716.196545] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 716.196545] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 716.196545] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] vm_ref = self.build_virtual_machine(instance, [ 716.197297] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 716.197297] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] vif_infos = vmwarevif.get_vif_info(self._session, [ 716.197297] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 716.197297] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] for vif in network_info: [ 716.197297] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 716.197297] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] return self._sync_wrapper(fn, *args, **kwargs) [ 716.197297] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 716.197297] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] self.wait() [ 716.197297] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 716.197297] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] self[:] = self._gt.wait() [ 716.197297] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 716.197297] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] return self._exit_event.wait() [ 716.197297] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 716.197738] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] result = hub.switch() [ 716.197738] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 716.197738] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] return self.greenlet.switch() [ 716.197738] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.197738] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] result = function(*args, **kwargs) [ 716.197738] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 716.197738] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] return func(*args, **kwargs) [ 716.197738] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 716.197738] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] raise e [ 716.197738] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.197738] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] nwinfo = self.network_api.allocate_for_instance( [ 716.197738] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 716.197738] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] created_port_ids = self._update_ports_for_instance( [ 716.198119] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 716.198119] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] with excutils.save_and_reraise_exception(): [ 716.198119] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.198119] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] self.force_reraise() [ 716.198119] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.198119] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] raise self.value [ 716.198119] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 716.198119] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] updated_port = self._update_port( [ 716.198119] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.198119] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] _ensure_no_port_binding_failure(port) [ 716.198119] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.198119] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] raise exception.PortBindingFailed(port_id=port['id']) [ 716.198450] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] nova.exception.PortBindingFailed: Binding failed for port ebd40d4d-5c50-4ebd-81a4-44ec4d9bfb2e, please check neutron logs for more information. [ 716.198450] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] [ 716.198450] env[61906]: INFO nova.compute.manager [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Terminating instance [ 716.200419] env[61906]: DEBUG oslo_concurrency.lockutils [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Acquiring lock "refresh_cache-9e665b58-69e7-4ab9-b109-ab27725e66cc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.200587] env[61906]: DEBUG oslo_concurrency.lockutils [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Acquired lock "refresh_cache-9e665b58-69e7-4ab9-b109-ab27725e66cc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.200775] env[61906]: DEBUG nova.network.neutron [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 716.486062] env[61906]: DEBUG nova.network.neutron [-] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.573387] env[61906]: DEBUG oslo_concurrency.lockutils [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.510s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.573910] env[61906]: DEBUG nova.compute.manager [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 716.577558] env[61906]: DEBUG oslo_concurrency.lockutils [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 20.648s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.577558] env[61906]: DEBUG nova.objects.instance [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61906) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 716.732486] env[61906]: DEBUG nova.network.neutron [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 716.851357] env[61906]: DEBUG nova.network.neutron [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.990567] env[61906]: INFO nova.compute.manager [-] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Took 1.03 seconds to deallocate network for instance. [ 716.993032] env[61906]: DEBUG nova.compute.claims [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 716.993187] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.081362] env[61906]: DEBUG nova.compute.utils [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 717.089119] env[61906]: DEBUG nova.compute.manager [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 717.089119] env[61906]: DEBUG nova.network.neutron [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 717.175152] env[61906]: DEBUG nova.policy [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '90fe1799f7df4e328e553f8b9906ef12', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3693c14b331e41ef80a38bd0a6318391', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 717.355967] env[61906]: DEBUG oslo_concurrency.lockutils [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Releasing lock "refresh_cache-9e665b58-69e7-4ab9-b109-ab27725e66cc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.356426] env[61906]: DEBUG nova.compute.manager [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 717.356619] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 717.356907] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e5751525-a8fc-4bad-afc9-03c2233678b5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.366893] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-964eca18-368f-4544-9fc9-ce16e876b0a1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.391333] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9e665b58-69e7-4ab9-b109-ab27725e66cc could not be found. [ 717.391564] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 717.391751] env[61906]: INFO nova.compute.manager [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Took 0.04 seconds to destroy the instance on the hypervisor. [ 717.392038] env[61906]: DEBUG oslo.service.loopingcall [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 717.392264] env[61906]: DEBUG nova.compute.manager [-] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 717.392353] env[61906]: DEBUG nova.network.neutron [-] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 717.408372] env[61906]: DEBUG nova.network.neutron [-] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 717.522237] env[61906]: DEBUG nova.network.neutron [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Successfully created port: 0ccf6630-3159-48b7-8f73-0112b29176eb {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 717.587690] env[61906]: DEBUG nova.compute.manager [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 717.591532] env[61906]: DEBUG oslo_concurrency.lockutils [None req-00855ed1-8071-4cea-99ce-75f830935bbd tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.592614] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.623s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.592828] env[61906]: DEBUG nova.objects.instance [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Lazy-loading 'resources' on Instance uuid 410b0308-1dca-4d11-91ab-f6d89bd565e3 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 717.843594] env[61906]: DEBUG nova.compute.manager [req-082cfc7a-b710-4f90-beab-b741d012e83f req-b98e4ea3-ea06-4db3-9e8b-2d3458279140 service nova] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Received event network-changed-ebd40d4d-5c50-4ebd-81a4-44ec4d9bfb2e {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 717.843817] env[61906]: DEBUG nova.compute.manager [req-082cfc7a-b710-4f90-beab-b741d012e83f req-b98e4ea3-ea06-4db3-9e8b-2d3458279140 service nova] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Refreshing instance network info cache due to event network-changed-ebd40d4d-5c50-4ebd-81a4-44ec4d9bfb2e. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 717.844069] env[61906]: DEBUG oslo_concurrency.lockutils [req-082cfc7a-b710-4f90-beab-b741d012e83f req-b98e4ea3-ea06-4db3-9e8b-2d3458279140 service nova] Acquiring lock "refresh_cache-9e665b58-69e7-4ab9-b109-ab27725e66cc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.844217] env[61906]: DEBUG oslo_concurrency.lockutils [req-082cfc7a-b710-4f90-beab-b741d012e83f req-b98e4ea3-ea06-4db3-9e8b-2d3458279140 service nova] Acquired lock "refresh_cache-9e665b58-69e7-4ab9-b109-ab27725e66cc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.844375] env[61906]: DEBUG nova.network.neutron [req-082cfc7a-b710-4f90-beab-b741d012e83f req-b98e4ea3-ea06-4db3-9e8b-2d3458279140 service nova] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Refreshing network info cache for port ebd40d4d-5c50-4ebd-81a4-44ec4d9bfb2e {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 717.910419] env[61906]: DEBUG nova.network.neutron [-] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.363968] env[61906]: DEBUG nova.network.neutron [req-082cfc7a-b710-4f90-beab-b741d012e83f req-b98e4ea3-ea06-4db3-9e8b-2d3458279140 service nova] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.414778] env[61906]: INFO nova.compute.manager [-] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Took 1.02 seconds to deallocate network for instance. [ 718.421120] env[61906]: DEBUG nova.compute.claims [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 718.421446] env[61906]: DEBUG oslo_concurrency.lockutils [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.475518] env[61906]: DEBUG nova.network.neutron [req-082cfc7a-b710-4f90-beab-b741d012e83f req-b98e4ea3-ea06-4db3-9e8b-2d3458279140 service nova] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.503065] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d49a931a-2237-44b4-b640-d8b4e514c5b0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.510922] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b871da1a-738e-4d6b-8637-dbee71fc30ee {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.544987] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3087c3ac-e04a-4c3d-a13d-d87e5bd6b96d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.552507] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95415151-8658-45e8-b7f8-0ce33056927b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.565980] env[61906]: DEBUG nova.compute.provider_tree [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 718.602706] env[61906]: DEBUG nova.compute.manager [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 718.628051] env[61906]: DEBUG nova.virt.hardware [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 718.628463] env[61906]: DEBUG nova.virt.hardware [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 718.628688] env[61906]: DEBUG nova.virt.hardware [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 718.628916] env[61906]: DEBUG nova.virt.hardware [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 718.629126] env[61906]: DEBUG nova.virt.hardware [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 718.629284] env[61906]: DEBUG nova.virt.hardware [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 718.629521] env[61906]: DEBUG nova.virt.hardware [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 718.629792] env[61906]: DEBUG nova.virt.hardware [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 718.630158] env[61906]: DEBUG nova.virt.hardware [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 718.630494] env[61906]: DEBUG nova.virt.hardware [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 718.630838] env[61906]: DEBUG nova.virt.hardware [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 718.631801] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-608a2121-3889-4e04-b70a-5061a635c3d8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.640774] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83f191f0-8c39-4937-b0b5-0eb26ab34c77 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.660571] env[61906]: ERROR nova.compute.manager [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0ccf6630-3159-48b7-8f73-0112b29176eb, please check neutron logs for more information. [ 718.660571] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 718.660571] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.660571] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 718.660571] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.660571] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 718.660571] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.660571] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 718.660571] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.660571] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 718.660571] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.660571] env[61906]: ERROR nova.compute.manager raise self.value [ 718.660571] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.660571] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 718.660571] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.660571] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 718.661169] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.661169] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 718.661169] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0ccf6630-3159-48b7-8f73-0112b29176eb, please check neutron logs for more information. [ 718.661169] env[61906]: ERROR nova.compute.manager [ 718.661169] env[61906]: Traceback (most recent call last): [ 718.661169] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 718.661169] env[61906]: listener.cb(fileno) [ 718.661169] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.661169] env[61906]: result = function(*args, **kwargs) [ 718.661169] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 718.661169] env[61906]: return func(*args, **kwargs) [ 718.661169] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 718.661169] env[61906]: raise e [ 718.661169] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.661169] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 718.661169] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.661169] env[61906]: created_port_ids = self._update_ports_for_instance( [ 718.661169] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.661169] env[61906]: with excutils.save_and_reraise_exception(): [ 718.661169] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.661169] env[61906]: self.force_reraise() [ 718.661169] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.661169] env[61906]: raise self.value [ 718.661169] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.661169] env[61906]: updated_port = self._update_port( [ 718.661169] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.661169] env[61906]: _ensure_no_port_binding_failure(port) [ 718.661169] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.661169] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 718.662057] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 0ccf6630-3159-48b7-8f73-0112b29176eb, please check neutron logs for more information. [ 718.662057] env[61906]: Removing descriptor: 18 [ 718.662057] env[61906]: ERROR nova.compute.manager [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0ccf6630-3159-48b7-8f73-0112b29176eb, please check neutron logs for more information. [ 718.662057] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Traceback (most recent call last): [ 718.662057] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 718.662057] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] yield resources [ 718.662057] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 718.662057] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] self.driver.spawn(context, instance, image_meta, [ 718.662057] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 718.662057] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] self._vmops.spawn(context, instance, image_meta, injected_files, [ 718.662057] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 718.662057] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] vm_ref = self.build_virtual_machine(instance, [ 718.662428] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 718.662428] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] vif_infos = vmwarevif.get_vif_info(self._session, [ 718.662428] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 718.662428] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] for vif in network_info: [ 718.662428] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 718.662428] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] return self._sync_wrapper(fn, *args, **kwargs) [ 718.662428] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 718.662428] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] self.wait() [ 718.662428] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 718.662428] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] self[:] = self._gt.wait() [ 718.662428] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 718.662428] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] return self._exit_event.wait() [ 718.662428] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 718.662820] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] result = hub.switch() [ 718.662820] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 718.662820] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] return self.greenlet.switch() [ 718.662820] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.662820] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] result = function(*args, **kwargs) [ 718.662820] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 718.662820] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] return func(*args, **kwargs) [ 718.662820] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 718.662820] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] raise e [ 718.662820] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.662820] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] nwinfo = self.network_api.allocate_for_instance( [ 718.662820] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.662820] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] created_port_ids = self._update_ports_for_instance( [ 718.663215] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.663215] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] with excutils.save_and_reraise_exception(): [ 718.663215] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.663215] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] self.force_reraise() [ 718.663215] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.663215] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] raise self.value [ 718.663215] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.663215] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] updated_port = self._update_port( [ 718.663215] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.663215] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] _ensure_no_port_binding_failure(port) [ 718.663215] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.663215] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] raise exception.PortBindingFailed(port_id=port['id']) [ 718.663595] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] nova.exception.PortBindingFailed: Binding failed for port 0ccf6630-3159-48b7-8f73-0112b29176eb, please check neutron logs for more information. [ 718.663595] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] [ 718.663595] env[61906]: INFO nova.compute.manager [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Terminating instance [ 718.668832] env[61906]: DEBUG oslo_concurrency.lockutils [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Acquiring lock "refresh_cache-8d2459c0-279c-4ea4-8cbc-6c69cfa35b27" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.669013] env[61906]: DEBUG oslo_concurrency.lockutils [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Acquired lock "refresh_cache-8d2459c0-279c-4ea4-8cbc-6c69cfa35b27" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.669188] env[61906]: DEBUG nova.network.neutron [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 718.980823] env[61906]: DEBUG oslo_concurrency.lockutils [req-082cfc7a-b710-4f90-beab-b741d012e83f req-b98e4ea3-ea06-4db3-9e8b-2d3458279140 service nova] Releasing lock "refresh_cache-9e665b58-69e7-4ab9-b109-ab27725e66cc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.980981] env[61906]: DEBUG nova.compute.manager [req-082cfc7a-b710-4f90-beab-b741d012e83f req-b98e4ea3-ea06-4db3-9e8b-2d3458279140 service nova] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Received event network-vif-deleted-ebd40d4d-5c50-4ebd-81a4-44ec4d9bfb2e {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 719.069252] env[61906]: DEBUG nova.scheduler.client.report [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 719.193127] env[61906]: DEBUG nova.network.neutron [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.264538] env[61906]: DEBUG nova.network.neutron [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.542157] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Acquiring lock "34f333be-ea19-499e-ab89-e9f7663ae596" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.542364] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Lock "34f333be-ea19-499e-ab89-e9f7663ae596" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.574479] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.982s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.577353] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.628s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.595951] env[61906]: INFO nova.scheduler.client.report [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Deleted allocations for instance 410b0308-1dca-4d11-91ab-f6d89bd565e3 [ 719.767020] env[61906]: DEBUG oslo_concurrency.lockutils [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Releasing lock "refresh_cache-8d2459c0-279c-4ea4-8cbc-6c69cfa35b27" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.767508] env[61906]: DEBUG nova.compute.manager [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 719.767740] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 719.768076] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1428db55-3298-4c17-b76c-26d53d1adf18 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.779432] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e13f2a3a-8a15-4511-80a5-baa8dee660b2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.801613] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27 could not be found. [ 719.801827] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 719.802017] env[61906]: INFO nova.compute.manager [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Took 0.03 seconds to destroy the instance on the hypervisor. [ 719.802270] env[61906]: DEBUG oslo.service.loopingcall [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 719.802485] env[61906]: DEBUG nova.compute.manager [-] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 719.802576] env[61906]: DEBUG nova.network.neutron [-] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 719.825206] env[61906]: DEBUG nova.network.neutron [-] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.869559] env[61906]: DEBUG nova.compute.manager [req-6d65716d-9a1f-44d0-9612-698d845d0168 req-434ba22e-821a-4cc9-ad57-fe8af6e816e8 service nova] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Received event network-changed-0ccf6630-3159-48b7-8f73-0112b29176eb {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 719.869748] env[61906]: DEBUG nova.compute.manager [req-6d65716d-9a1f-44d0-9612-698d845d0168 req-434ba22e-821a-4cc9-ad57-fe8af6e816e8 service nova] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Refreshing instance network info cache due to event network-changed-0ccf6630-3159-48b7-8f73-0112b29176eb. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 719.869961] env[61906]: DEBUG oslo_concurrency.lockutils [req-6d65716d-9a1f-44d0-9612-698d845d0168 req-434ba22e-821a-4cc9-ad57-fe8af6e816e8 service nova] Acquiring lock "refresh_cache-8d2459c0-279c-4ea4-8cbc-6c69cfa35b27" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.870118] env[61906]: DEBUG oslo_concurrency.lockutils [req-6d65716d-9a1f-44d0-9612-698d845d0168 req-434ba22e-821a-4cc9-ad57-fe8af6e816e8 service nova] Acquired lock "refresh_cache-8d2459c0-279c-4ea4-8cbc-6c69cfa35b27" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.870278] env[61906]: DEBUG nova.network.neutron [req-6d65716d-9a1f-44d0-9612-698d845d0168 req-434ba22e-821a-4cc9-ad57-fe8af6e816e8 service nova] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Refreshing network info cache for port 0ccf6630-3159-48b7-8f73-0112b29176eb {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 720.105884] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a3a9c8d4-2b39-4214-bf47-8a281548008a tempest-ServerShowV254Test-998596363 tempest-ServerShowV254Test-998596363-project-member] Lock "410b0308-1dca-4d11-91ab-f6d89bd565e3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.842s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.328722] env[61906]: DEBUG nova.network.neutron [-] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.388902] env[61906]: DEBUG nova.network.neutron [req-6d65716d-9a1f-44d0-9612-698d845d0168 req-434ba22e-821a-4cc9-ad57-fe8af6e816e8 service nova] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.394717] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b48941ce-62eb-4e7e-9a5d-079789fa9aa2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.403436] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3312222f-6e97-42a0-9653-35313893eef8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.434190] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-442f91a2-136d-4ea9-b460-9cd78f3d654d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.442376] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f39d4b1a-c190-47e1-a0e7-fc598c74a719 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.457443] env[61906]: DEBUG nova.compute.provider_tree [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.511940] env[61906]: DEBUG nova.network.neutron [req-6d65716d-9a1f-44d0-9612-698d845d0168 req-434ba22e-821a-4cc9-ad57-fe8af6e816e8 service nova] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.831033] env[61906]: INFO nova.compute.manager [-] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Took 1.03 seconds to deallocate network for instance. [ 720.833499] env[61906]: DEBUG nova.compute.claims [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 720.833674] env[61906]: DEBUG oslo_concurrency.lockutils [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.963622] env[61906]: DEBUG nova.scheduler.client.report [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 721.015024] env[61906]: DEBUG oslo_concurrency.lockutils [req-6d65716d-9a1f-44d0-9612-698d845d0168 req-434ba22e-821a-4cc9-ad57-fe8af6e816e8 service nova] Releasing lock "refresh_cache-8d2459c0-279c-4ea4-8cbc-6c69cfa35b27" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.015024] env[61906]: DEBUG nova.compute.manager [req-6d65716d-9a1f-44d0-9612-698d845d0168 req-434ba22e-821a-4cc9-ad57-fe8af6e816e8 service nova] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Received event network-vif-deleted-0ccf6630-3159-48b7-8f73-0112b29176eb {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 721.469944] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.893s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.470597] env[61906]: ERROR nova.compute.manager [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 99389367-21dd-45ae-8f2e-c1002dc479dd, please check neutron logs for more information. [ 721.470597] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Traceback (most recent call last): [ 721.470597] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 721.470597] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] self.driver.spawn(context, instance, image_meta, [ 721.470597] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 721.470597] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 721.470597] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 721.470597] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] vm_ref = self.build_virtual_machine(instance, [ 721.470597] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 721.470597] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] vif_infos = vmwarevif.get_vif_info(self._session, [ 721.470597] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 721.470981] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] for vif in network_info: [ 721.470981] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 721.470981] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] return self._sync_wrapper(fn, *args, **kwargs) [ 721.470981] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 721.470981] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] self.wait() [ 721.470981] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 721.470981] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] self[:] = self._gt.wait() [ 721.470981] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 721.470981] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] return self._exit_event.wait() [ 721.470981] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 721.470981] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] current.throw(*self._exc) [ 721.470981] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 721.470981] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] result = function(*args, **kwargs) [ 721.471380] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 721.471380] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] return func(*args, **kwargs) [ 721.471380] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 721.471380] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] raise e [ 721.471380] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.471380] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] nwinfo = self.network_api.allocate_for_instance( [ 721.471380] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.471380] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] created_port_ids = self._update_ports_for_instance( [ 721.471380] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.471380] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] with excutils.save_and_reraise_exception(): [ 721.471380] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.471380] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] self.force_reraise() [ 721.471380] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.471724] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] raise self.value [ 721.471724] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.471724] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] updated_port = self._update_port( [ 721.471724] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.471724] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] _ensure_no_port_binding_failure(port) [ 721.471724] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.471724] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] raise exception.PortBindingFailed(port_id=port['id']) [ 721.471724] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] nova.exception.PortBindingFailed: Binding failed for port 99389367-21dd-45ae-8f2e-c1002dc479dd, please check neutron logs for more information. [ 721.471724] env[61906]: ERROR nova.compute.manager [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] [ 721.471724] env[61906]: DEBUG nova.compute.utils [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Binding failed for port 99389367-21dd-45ae-8f2e-c1002dc479dd, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 721.472549] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.414s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.479795] env[61906]: INFO nova.compute.claims [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 721.482628] env[61906]: DEBUG nova.compute.manager [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Build of instance 2e8d82fd-5f70-408e-9d3b-74f3d81232d3 was re-scheduled: Binding failed for port 99389367-21dd-45ae-8f2e-c1002dc479dd, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 721.483113] env[61906]: DEBUG nova.compute.manager [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 721.483359] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Acquiring lock "refresh_cache-2e8d82fd-5f70-408e-9d3b-74f3d81232d3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.483505] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Acquired lock "refresh_cache-2e8d82fd-5f70-408e-9d3b-74f3d81232d3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.483664] env[61906]: DEBUG nova.network.neutron [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 722.006057] env[61906]: DEBUG nova.network.neutron [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.094775] env[61906]: DEBUG nova.network.neutron [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.596816] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Releasing lock "refresh_cache-2e8d82fd-5f70-408e-9d3b-74f3d81232d3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.597065] env[61906]: DEBUG nova.compute.manager [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 722.597251] env[61906]: DEBUG nova.compute.manager [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 722.597427] env[61906]: DEBUG nova.network.neutron [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 722.617163] env[61906]: DEBUG nova.network.neutron [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.782370] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-522462dc-5fb1-4c91-a687-bcf922bb464a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.789794] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d433b1f4-f812-4b8a-b814-76af6dceebc6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.818799] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea22ba02-1ce8-4d12-8b2f-988b9bebeca1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.825980] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37ec0448-adc5-486f-a913-1a0023be3eb0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.840347] env[61906]: DEBUG nova.compute.provider_tree [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.121727] env[61906]: DEBUG nova.network.neutron [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.343692] env[61906]: DEBUG nova.scheduler.client.report [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 723.624847] env[61906]: INFO nova.compute.manager [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] [instance: 2e8d82fd-5f70-408e-9d3b-74f3d81232d3] Took 1.03 seconds to deallocate network for instance. [ 723.849213] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.376s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.849728] env[61906]: DEBUG nova.compute.manager [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 723.852492] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.861s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.853709] env[61906]: INFO nova.compute.claims [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 724.358230] env[61906]: DEBUG nova.compute.utils [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 724.361948] env[61906]: DEBUG nova.compute.manager [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 724.361948] env[61906]: DEBUG nova.network.neutron [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 724.403133] env[61906]: DEBUG nova.policy [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e883fb94e9b74e54925677b3f9466e26', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c15bc344a5440d491b47b9fdedb6869', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 724.665769] env[61906]: INFO nova.scheduler.client.report [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Deleted allocations for instance 2e8d82fd-5f70-408e-9d3b-74f3d81232d3 [ 724.675298] env[61906]: DEBUG nova.network.neutron [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Successfully created port: 2281c028-9a73-46e9-be06-40aeccecd788 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 724.864036] env[61906]: DEBUG nova.compute.manager [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 724.962185] env[61906]: DEBUG nova.network.neutron [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Successfully created port: 1cba8868-7c8b-4aea-9e2b-610c1c105cbe {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 725.175082] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f8267c0-8cbe-4b68-8e70-e78e7a76e7dd tempest-ServerActionsV293TestJSON-1265403277 tempest-ServerActionsV293TestJSON-1265403277-project-member] Lock "2e8d82fd-5f70-408e-9d3b-74f3d81232d3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.648s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.195434] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ffcacc1-8c6b-4fc4-8eb2-f0c4a168a117 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.206684] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d8fff7-634a-4c99-9792-0840ace728a0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.241608] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41d5ed70-c110-495e-a6fa-55c7bbdbd844 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.250154] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb9ddcee-fbf0-4a3d-a47b-36f61253cdd1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.255220] env[61906]: DEBUG nova.network.neutron [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Successfully created port: 82d3cad7-db7c-481a-95d5-ff7924661774 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 725.269420] env[61906]: DEBUG nova.compute.provider_tree [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.676817] env[61906]: DEBUG nova.compute.manager [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 725.773138] env[61906]: DEBUG nova.scheduler.client.report [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 725.874157] env[61906]: DEBUG nova.compute.manager [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 725.908315] env[61906]: DEBUG nova.virt.hardware [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 725.908622] env[61906]: DEBUG nova.virt.hardware [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 725.908874] env[61906]: DEBUG nova.virt.hardware [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 725.909159] env[61906]: DEBUG nova.virt.hardware [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 725.909356] env[61906]: DEBUG nova.virt.hardware [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 725.909550] env[61906]: DEBUG nova.virt.hardware [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 725.909815] env[61906]: DEBUG nova.virt.hardware [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 725.909990] env[61906]: DEBUG nova.virt.hardware [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 725.910174] env[61906]: DEBUG nova.virt.hardware [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 725.910339] env[61906]: DEBUG nova.virt.hardware [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 725.910508] env[61906]: DEBUG nova.virt.hardware [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 725.911411] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e08f762d-34fa-4145-a6b8-746fe9302360 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.920816] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35fbfbc0-80a7-4375-8a57-6a216dfb59cd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.203351] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.243683] env[61906]: DEBUG nova.compute.manager [req-c768d2f2-4ac2-4d5d-8fe8-2954d5b286cd req-859f6383-2485-4298-b33a-957a32a94727 service nova] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Received event network-changed-2281c028-9a73-46e9-be06-40aeccecd788 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 726.243886] env[61906]: DEBUG nova.compute.manager [req-c768d2f2-4ac2-4d5d-8fe8-2954d5b286cd req-859f6383-2485-4298-b33a-957a32a94727 service nova] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Refreshing instance network info cache due to event network-changed-2281c028-9a73-46e9-be06-40aeccecd788. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 726.244115] env[61906]: DEBUG oslo_concurrency.lockutils [req-c768d2f2-4ac2-4d5d-8fe8-2954d5b286cd req-859f6383-2485-4298-b33a-957a32a94727 service nova] Acquiring lock "refresh_cache-7b4a9ebc-47fb-4078-969a-07642284b916" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.244275] env[61906]: DEBUG oslo_concurrency.lockutils [req-c768d2f2-4ac2-4d5d-8fe8-2954d5b286cd req-859f6383-2485-4298-b33a-957a32a94727 service nova] Acquired lock "refresh_cache-7b4a9ebc-47fb-4078-969a-07642284b916" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.244410] env[61906]: DEBUG nova.network.neutron [req-c768d2f2-4ac2-4d5d-8fe8-2954d5b286cd req-859f6383-2485-4298-b33a-957a32a94727 service nova] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Refreshing network info cache for port 2281c028-9a73-46e9-be06-40aeccecd788 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 726.277891] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.425s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.278392] env[61906]: DEBUG nova.compute.manager [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 726.282715] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.857s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.454125] env[61906]: ERROR nova.compute.manager [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2281c028-9a73-46e9-be06-40aeccecd788, please check neutron logs for more information. [ 726.454125] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 726.454125] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.454125] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 726.454125] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.454125] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 726.454125] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.454125] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 726.454125] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.454125] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 726.454125] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.454125] env[61906]: ERROR nova.compute.manager raise self.value [ 726.454125] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.454125] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 726.454125] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.454125] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 726.454665] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.454665] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 726.454665] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2281c028-9a73-46e9-be06-40aeccecd788, please check neutron logs for more information. [ 726.454665] env[61906]: ERROR nova.compute.manager [ 726.454665] env[61906]: Traceback (most recent call last): [ 726.454665] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 726.454665] env[61906]: listener.cb(fileno) [ 726.454665] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.454665] env[61906]: result = function(*args, **kwargs) [ 726.454665] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 726.454665] env[61906]: return func(*args, **kwargs) [ 726.454665] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 726.454665] env[61906]: raise e [ 726.454665] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.454665] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 726.454665] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.454665] env[61906]: created_port_ids = self._update_ports_for_instance( [ 726.454665] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.454665] env[61906]: with excutils.save_and_reraise_exception(): [ 726.454665] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.454665] env[61906]: self.force_reraise() [ 726.454665] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.454665] env[61906]: raise self.value [ 726.454665] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.454665] env[61906]: updated_port = self._update_port( [ 726.454665] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.454665] env[61906]: _ensure_no_port_binding_failure(port) [ 726.454665] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.454665] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 726.455642] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 2281c028-9a73-46e9-be06-40aeccecd788, please check neutron logs for more information. [ 726.455642] env[61906]: Removing descriptor: 18 [ 726.455642] env[61906]: ERROR nova.compute.manager [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2281c028-9a73-46e9-be06-40aeccecd788, please check neutron logs for more information. [ 726.455642] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Traceback (most recent call last): [ 726.455642] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 726.455642] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] yield resources [ 726.455642] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 726.455642] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] self.driver.spawn(context, instance, image_meta, [ 726.455642] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 726.455642] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] self._vmops.spawn(context, instance, image_meta, injected_files, [ 726.455642] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 726.455642] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] vm_ref = self.build_virtual_machine(instance, [ 726.456051] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 726.456051] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] vif_infos = vmwarevif.get_vif_info(self._session, [ 726.456051] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 726.456051] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] for vif in network_info: [ 726.456051] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 726.456051] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] return self._sync_wrapper(fn, *args, **kwargs) [ 726.456051] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 726.456051] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] self.wait() [ 726.456051] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 726.456051] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] self[:] = self._gt.wait() [ 726.456051] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 726.456051] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] return self._exit_event.wait() [ 726.456051] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 726.456434] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] result = hub.switch() [ 726.456434] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 726.456434] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] return self.greenlet.switch() [ 726.456434] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.456434] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] result = function(*args, **kwargs) [ 726.456434] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 726.456434] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] return func(*args, **kwargs) [ 726.456434] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 726.456434] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] raise e [ 726.456434] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.456434] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] nwinfo = self.network_api.allocate_for_instance( [ 726.456434] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.456434] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] created_port_ids = self._update_ports_for_instance( [ 726.456849] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.456849] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] with excutils.save_and_reraise_exception(): [ 726.456849] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.456849] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] self.force_reraise() [ 726.456849] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.456849] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] raise self.value [ 726.456849] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.456849] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] updated_port = self._update_port( [ 726.456849] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.456849] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] _ensure_no_port_binding_failure(port) [ 726.456849] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.456849] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] raise exception.PortBindingFailed(port_id=port['id']) [ 726.457231] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] nova.exception.PortBindingFailed: Binding failed for port 2281c028-9a73-46e9-be06-40aeccecd788, please check neutron logs for more information. [ 726.457231] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] [ 726.457231] env[61906]: INFO nova.compute.manager [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Terminating instance [ 726.459665] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Acquiring lock "refresh_cache-7b4a9ebc-47fb-4078-969a-07642284b916" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.786027] env[61906]: DEBUG nova.compute.utils [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 726.787817] env[61906]: DEBUG nova.compute.manager [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 726.787976] env[61906]: DEBUG nova.network.neutron [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 726.823218] env[61906]: DEBUG nova.network.neutron [req-c768d2f2-4ac2-4d5d-8fe8-2954d5b286cd req-859f6383-2485-4298-b33a-957a32a94727 service nova] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.869912] env[61906]: DEBUG nova.policy [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bfb8ebd8f94f45dfadb1d8802d04aec9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4f7fc2042e6549cabe98838990f35842', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 726.944026] env[61906]: DEBUG nova.network.neutron [req-c768d2f2-4ac2-4d5d-8fe8-2954d5b286cd req-859f6383-2485-4298-b33a-957a32a94727 service nova] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.173726] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c01c7b5-b0fc-4316-9fe3-c85826e95687 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.183469] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef0915bf-8ccf-44ba-b615-bd53c347cd58 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.214569] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fec00ae-b78d-47ab-bc73-0dc02d07a682 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.221891] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8c3fd0f-8d3c-4ffa-b5ab-16f760a8d1f3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.235862] env[61906]: DEBUG nova.compute.provider_tree [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.293936] env[61906]: DEBUG nova.compute.manager [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 727.356222] env[61906]: DEBUG nova.network.neutron [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Successfully created port: 6646962d-b310-45c7-924b-fcd70fcfe261 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 727.450544] env[61906]: DEBUG oslo_concurrency.lockutils [req-c768d2f2-4ac2-4d5d-8fe8-2954d5b286cd req-859f6383-2485-4298-b33a-957a32a94727 service nova] Releasing lock "refresh_cache-7b4a9ebc-47fb-4078-969a-07642284b916" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.452065] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Acquired lock "refresh_cache-7b4a9ebc-47fb-4078-969a-07642284b916" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.452065] env[61906]: DEBUG nova.network.neutron [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 727.739122] env[61906]: DEBUG nova.scheduler.client.report [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 728.005923] env[61906]: DEBUG nova.network.neutron [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.111871] env[61906]: DEBUG nova.network.neutron [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.217024] env[61906]: DEBUG nova.compute.manager [req-fbdcdcee-f82b-4c87-adc8-675166e58375 req-96f8588d-c838-4126-8d11-f96fc3807a0f service nova] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Received event network-changed-6646962d-b310-45c7-924b-fcd70fcfe261 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 728.217024] env[61906]: DEBUG nova.compute.manager [req-fbdcdcee-f82b-4c87-adc8-675166e58375 req-96f8588d-c838-4126-8d11-f96fc3807a0f service nova] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Refreshing instance network info cache due to event network-changed-6646962d-b310-45c7-924b-fcd70fcfe261. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 728.217024] env[61906]: DEBUG oslo_concurrency.lockutils [req-fbdcdcee-f82b-4c87-adc8-675166e58375 req-96f8588d-c838-4126-8d11-f96fc3807a0f service nova] Acquiring lock "refresh_cache-7c389914-47a6-4ed4-aad6-e8aeeefbed7c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.217024] env[61906]: DEBUG oslo_concurrency.lockutils [req-fbdcdcee-f82b-4c87-adc8-675166e58375 req-96f8588d-c838-4126-8d11-f96fc3807a0f service nova] Acquired lock "refresh_cache-7c389914-47a6-4ed4-aad6-e8aeeefbed7c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.217024] env[61906]: DEBUG nova.network.neutron [req-fbdcdcee-f82b-4c87-adc8-675166e58375 req-96f8588d-c838-4126-8d11-f96fc3807a0f service nova] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Refreshing network info cache for port 6646962d-b310-45c7-924b-fcd70fcfe261 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 728.243719] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.961s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.244421] env[61906]: ERROR nova.compute.manager [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5e472a5c-dd42-4934-a6ad-553423b8f024, please check neutron logs for more information. [ 728.244421] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Traceback (most recent call last): [ 728.244421] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 728.244421] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] self.driver.spawn(context, instance, image_meta, [ 728.244421] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 728.244421] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] self._vmops.spawn(context, instance, image_meta, injected_files, [ 728.244421] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 728.244421] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] vm_ref = self.build_virtual_machine(instance, [ 728.244421] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 728.244421] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] vif_infos = vmwarevif.get_vif_info(self._session, [ 728.244421] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 728.244993] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] for vif in network_info: [ 728.244993] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 728.244993] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] return self._sync_wrapper(fn, *args, **kwargs) [ 728.244993] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 728.244993] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] self.wait() [ 728.244993] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 728.244993] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] self[:] = self._gt.wait() [ 728.244993] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 728.244993] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] return self._exit_event.wait() [ 728.244993] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 728.244993] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] result = hub.switch() [ 728.244993] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 728.244993] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] return self.greenlet.switch() [ 728.246092] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.246092] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] result = function(*args, **kwargs) [ 728.246092] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 728.246092] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] return func(*args, **kwargs) [ 728.246092] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 728.246092] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] raise e [ 728.246092] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.246092] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] nwinfo = self.network_api.allocate_for_instance( [ 728.246092] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 728.246092] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] created_port_ids = self._update_ports_for_instance( [ 728.246092] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 728.246092] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] with excutils.save_and_reraise_exception(): [ 728.246092] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.246705] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] self.force_reraise() [ 728.246705] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.246705] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] raise self.value [ 728.246705] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 728.246705] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] updated_port = self._update_port( [ 728.246705] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.246705] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] _ensure_no_port_binding_failure(port) [ 728.246705] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.246705] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] raise exception.PortBindingFailed(port_id=port['id']) [ 728.246705] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] nova.exception.PortBindingFailed: Binding failed for port 5e472a5c-dd42-4934-a6ad-553423b8f024, please check neutron logs for more information. [ 728.246705] env[61906]: ERROR nova.compute.manager [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] [ 728.247326] env[61906]: DEBUG nova.compute.utils [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Binding failed for port 5e472a5c-dd42-4934-a6ad-553423b8f024, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 728.248840] env[61906]: DEBUG oslo_concurrency.lockutils [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.651s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.250295] env[61906]: INFO nova.compute.claims [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 728.253199] env[61906]: DEBUG nova.compute.manager [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Build of instance fd30cb23-f30f-4e20-ae67-36956b78e6ac was re-scheduled: Binding failed for port 5e472a5c-dd42-4934-a6ad-553423b8f024, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 728.253605] env[61906]: DEBUG nova.compute.manager [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 728.253829] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "refresh_cache-fd30cb23-f30f-4e20-ae67-36956b78e6ac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.253976] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquired lock "refresh_cache-fd30cb23-f30f-4e20-ae67-36956b78e6ac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.254150] env[61906]: DEBUG nova.network.neutron [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 728.270935] env[61906]: DEBUG nova.compute.manager [req-13e8b572-0b15-41fb-a6e2-ebfef1e6a8fa req-8c2f5e17-0f9c-4fc2-9ee9-c5f6af3eaaff service nova] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Received event network-vif-deleted-2281c028-9a73-46e9-be06-40aeccecd788 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 728.306248] env[61906]: DEBUG nova.compute.manager [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 728.333084] env[61906]: DEBUG nova.virt.hardware [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 728.333369] env[61906]: DEBUG nova.virt.hardware [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 728.333526] env[61906]: DEBUG nova.virt.hardware [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 728.333708] env[61906]: DEBUG nova.virt.hardware [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 728.333851] env[61906]: DEBUG nova.virt.hardware [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 728.333997] env[61906]: DEBUG nova.virt.hardware [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 728.334215] env[61906]: DEBUG nova.virt.hardware [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 728.334373] env[61906]: DEBUG nova.virt.hardware [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 728.334534] env[61906]: DEBUG nova.virt.hardware [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 728.334690] env[61906]: DEBUG nova.virt.hardware [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 728.334860] env[61906]: DEBUG nova.virt.hardware [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 728.335746] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55038b3f-4863-4895-8a7b-c598fb3b0a5a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.344222] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eefd9689-a112-4ec3-8c53-843d8105492c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.376475] env[61906]: ERROR nova.compute.manager [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6646962d-b310-45c7-924b-fcd70fcfe261, please check neutron logs for more information. [ 728.376475] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 728.376475] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.376475] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 728.376475] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 728.376475] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 728.376475] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 728.376475] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 728.376475] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.376475] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 728.376475] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.376475] env[61906]: ERROR nova.compute.manager raise self.value [ 728.376475] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 728.376475] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 728.376475] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.376475] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 728.377009] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.377009] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 728.377009] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6646962d-b310-45c7-924b-fcd70fcfe261, please check neutron logs for more information. [ 728.377009] env[61906]: ERROR nova.compute.manager [ 728.377009] env[61906]: Traceback (most recent call last): [ 728.377009] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 728.377009] env[61906]: listener.cb(fileno) [ 728.377009] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.377009] env[61906]: result = function(*args, **kwargs) [ 728.377009] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 728.377009] env[61906]: return func(*args, **kwargs) [ 728.377009] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 728.377009] env[61906]: raise e [ 728.377009] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.377009] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 728.377009] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 728.377009] env[61906]: created_port_ids = self._update_ports_for_instance( [ 728.377009] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 728.377009] env[61906]: with excutils.save_and_reraise_exception(): [ 728.377009] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.377009] env[61906]: self.force_reraise() [ 728.377009] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.377009] env[61906]: raise self.value [ 728.377009] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 728.377009] env[61906]: updated_port = self._update_port( [ 728.377009] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.377009] env[61906]: _ensure_no_port_binding_failure(port) [ 728.377009] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.377009] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 728.377856] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 6646962d-b310-45c7-924b-fcd70fcfe261, please check neutron logs for more information. [ 728.377856] env[61906]: Removing descriptor: 16 [ 728.377856] env[61906]: ERROR nova.compute.manager [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6646962d-b310-45c7-924b-fcd70fcfe261, please check neutron logs for more information. [ 728.377856] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Traceback (most recent call last): [ 728.377856] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 728.377856] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] yield resources [ 728.377856] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 728.377856] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] self.driver.spawn(context, instance, image_meta, [ 728.377856] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 728.377856] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 728.377856] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 728.377856] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] vm_ref = self.build_virtual_machine(instance, [ 728.378247] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 728.378247] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] vif_infos = vmwarevif.get_vif_info(self._session, [ 728.378247] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 728.378247] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] for vif in network_info: [ 728.378247] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 728.378247] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] return self._sync_wrapper(fn, *args, **kwargs) [ 728.378247] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 728.378247] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] self.wait() [ 728.378247] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 728.378247] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] self[:] = self._gt.wait() [ 728.378247] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 728.378247] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] return self._exit_event.wait() [ 728.378247] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 728.378765] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] result = hub.switch() [ 728.378765] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 728.378765] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] return self.greenlet.switch() [ 728.378765] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.378765] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] result = function(*args, **kwargs) [ 728.378765] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 728.378765] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] return func(*args, **kwargs) [ 728.378765] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 728.378765] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] raise e [ 728.378765] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.378765] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] nwinfo = self.network_api.allocate_for_instance( [ 728.378765] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 728.378765] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] created_port_ids = self._update_ports_for_instance( [ 728.379209] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 728.379209] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] with excutils.save_and_reraise_exception(): [ 728.379209] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.379209] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] self.force_reraise() [ 728.379209] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.379209] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] raise self.value [ 728.379209] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 728.379209] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] updated_port = self._update_port( [ 728.379209] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.379209] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] _ensure_no_port_binding_failure(port) [ 728.379209] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.379209] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] raise exception.PortBindingFailed(port_id=port['id']) [ 728.379577] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] nova.exception.PortBindingFailed: Binding failed for port 6646962d-b310-45c7-924b-fcd70fcfe261, please check neutron logs for more information. [ 728.379577] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] [ 728.379577] env[61906]: INFO nova.compute.manager [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Terminating instance [ 728.379693] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "refresh_cache-7c389914-47a6-4ed4-aad6-e8aeeefbed7c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.615073] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Releasing lock "refresh_cache-7b4a9ebc-47fb-4078-969a-07642284b916" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.615505] env[61906]: DEBUG nova.compute.manager [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 728.615695] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 728.615999] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1ec70c35-b674-4a19-876a-789c8bb2a016 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.624976] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d64b5eb-db7b-4139-ab55-cc6439ee7f3d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.646912] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7b4a9ebc-47fb-4078-969a-07642284b916 could not be found. [ 728.647138] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 728.647319] env[61906]: INFO nova.compute.manager [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Took 0.03 seconds to destroy the instance on the hypervisor. [ 728.647579] env[61906]: DEBUG oslo.service.loopingcall [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 728.647741] env[61906]: DEBUG nova.compute.manager [-] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 728.647836] env[61906]: DEBUG nova.network.neutron [-] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 728.706684] env[61906]: DEBUG nova.network.neutron [-] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.738143] env[61906]: DEBUG nova.network.neutron [req-fbdcdcee-f82b-4c87-adc8-675166e58375 req-96f8588d-c838-4126-8d11-f96fc3807a0f service nova] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.784635] env[61906]: DEBUG nova.network.neutron [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.897014] env[61906]: DEBUG nova.network.neutron [req-fbdcdcee-f82b-4c87-adc8-675166e58375 req-96f8588d-c838-4126-8d11-f96fc3807a0f service nova] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.902840] env[61906]: DEBUG nova.network.neutron [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.400088] env[61906]: DEBUG oslo_concurrency.lockutils [req-fbdcdcee-f82b-4c87-adc8-675166e58375 req-96f8588d-c838-4126-8d11-f96fc3807a0f service nova] Releasing lock "refresh_cache-7c389914-47a6-4ed4-aad6-e8aeeefbed7c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.402528] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired lock "refresh_cache-7c389914-47a6-4ed4-aad6-e8aeeefbed7c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.402717] env[61906]: DEBUG nova.network.neutron [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 729.404515] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Releasing lock "refresh_cache-fd30cb23-f30f-4e20-ae67-36956b78e6ac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.404621] env[61906]: DEBUG nova.compute.manager [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 729.404798] env[61906]: DEBUG nova.compute.manager [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 729.404957] env[61906]: DEBUG nova.network.neutron [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 729.424207] env[61906]: DEBUG nova.network.neutron [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.581139] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c616f6e1-086e-4744-9692-d4f914ed9b51 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.588771] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99cca8eb-6c8b-448a-8434-d6462175b3d3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.616919] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c387feb9-f34b-40e7-8564-e31c54c692a6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.623948] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24aa88e0-e062-42f1-8059-f63a2f7a6a90 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.637644] env[61906]: DEBUG nova.compute.provider_tree [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 729.735559] env[61906]: DEBUG nova.network.neutron [-] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.924123] env[61906]: DEBUG nova.network.neutron [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.926221] env[61906]: DEBUG nova.network.neutron [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.041059] env[61906]: DEBUG nova.network.neutron [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.141299] env[61906]: DEBUG nova.scheduler.client.report [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 730.239731] env[61906]: INFO nova.compute.manager [-] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Took 1.59 seconds to deallocate network for instance. [ 730.245158] env[61906]: DEBUG nova.compute.claims [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 730.245366] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.266590] env[61906]: DEBUG nova.compute.manager [req-2f8981c0-fa30-4037-bed6-b29cb6d5c221 req-3a039c4c-0ad5-4083-9c9a-d73f9ef6cec8 service nova] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Received event network-vif-deleted-6646962d-b310-45c7-924b-fcd70fcfe261 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 730.428724] env[61906]: INFO nova.compute.manager [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: fd30cb23-f30f-4e20-ae67-36956b78e6ac] Took 1.02 seconds to deallocate network for instance. [ 730.548534] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Releasing lock "refresh_cache-7c389914-47a6-4ed4-aad6-e8aeeefbed7c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.549148] env[61906]: DEBUG nova.compute.manager [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 730.549282] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 730.549511] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-743ad761-2d28-4ad7-b783-5cd5e8bf6515 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.559396] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04187fbc-2ee3-471e-83e1-3b3e289f9b36 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.584550] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7c389914-47a6-4ed4-aad6-e8aeeefbed7c could not be found. [ 730.585441] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 730.585441] env[61906]: INFO nova.compute.manager [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 730.585737] env[61906]: DEBUG oslo.service.loopingcall [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 730.588954] env[61906]: DEBUG nova.compute.manager [-] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 730.588954] env[61906]: DEBUG nova.network.neutron [-] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 730.603971] env[61906]: DEBUG nova.network.neutron [-] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.650944] env[61906]: DEBUG oslo_concurrency.lockutils [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.404s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.651565] env[61906]: DEBUG nova.compute.manager [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 730.654116] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.064s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.656108] env[61906]: INFO nova.compute.claims [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 731.107303] env[61906]: DEBUG nova.network.neutron [-] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.161321] env[61906]: DEBUG nova.compute.utils [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 731.165665] env[61906]: DEBUG nova.compute.manager [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 731.165902] env[61906]: DEBUG nova.network.neutron [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 731.228915] env[61906]: DEBUG nova.policy [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '463262c2e40e4db4a4469aaa5e6ac708', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '03f866963a584575a4444b1f0493cd71', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 731.469277] env[61906]: INFO nova.scheduler.client.report [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Deleted allocations for instance fd30cb23-f30f-4e20-ae67-36956b78e6ac [ 731.572197] env[61906]: DEBUG nova.network.neutron [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Successfully created port: 8aa8e05d-7e39-402e-b0a1-d5ca5fc4e630 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 731.612765] env[61906]: INFO nova.compute.manager [-] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Took 1.03 seconds to deallocate network for instance. [ 731.614377] env[61906]: DEBUG nova.compute.claims [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 731.614553] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.666949] env[61906]: DEBUG nova.compute.manager [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 731.758255] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Acquiring lock "7183e141-9989-4a90-b2fb-2ca0d8e6cda9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.758500] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Lock "7183e141-9989-4a90-b2fb-2ca0d8e6cda9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.978023] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8eb54be7-981f-4c93-970a-c781358b60bf tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "fd30cb23-f30f-4e20-ae67-36956b78e6ac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.157s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.071191] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bca1685-8384-4edf-aea7-f12383b4847c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.078804] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72eab5a3-b697-4823-a987-51c2ad78935e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.109086] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9d6a229-c5be-42b9-9f24-a3eb170aa8b6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.116590] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c5496e2-a400-4077-8c24-d2036658f91a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.130870] env[61906]: DEBUG nova.compute.provider_tree [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 732.481092] env[61906]: DEBUG nova.compute.manager [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 732.635161] env[61906]: DEBUG nova.scheduler.client.report [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 732.679135] env[61906]: DEBUG nova.compute.manager [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 732.708458] env[61906]: DEBUG nova.virt.hardware [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 732.708701] env[61906]: DEBUG nova.virt.hardware [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 732.708878] env[61906]: DEBUG nova.virt.hardware [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 732.709097] env[61906]: DEBUG nova.virt.hardware [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 732.709441] env[61906]: DEBUG nova.virt.hardware [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 732.709518] env[61906]: DEBUG nova.virt.hardware [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 732.709706] env[61906]: DEBUG nova.virt.hardware [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 732.709890] env[61906]: DEBUG nova.virt.hardware [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 732.710026] env[61906]: DEBUG nova.virt.hardware [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 732.710412] env[61906]: DEBUG nova.virt.hardware [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 732.710412] env[61906]: DEBUG nova.virt.hardware [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 732.711290] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e06a0d9-5d83-4e1a-9ed9-42d98368ffdd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.719311] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-969e82c7-8235-40ef-9405-d25938b89e8c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.830801] env[61906]: DEBUG nova.compute.manager [req-5fe76ab6-68ee-4e4a-8fdd-8a3fe18b5125 req-0219be2c-3457-4b31-9283-f68592a323d9 service nova] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Received event network-changed-8aa8e05d-7e39-402e-b0a1-d5ca5fc4e630 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 732.831102] env[61906]: DEBUG nova.compute.manager [req-5fe76ab6-68ee-4e4a-8fdd-8a3fe18b5125 req-0219be2c-3457-4b31-9283-f68592a323d9 service nova] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Refreshing instance network info cache due to event network-changed-8aa8e05d-7e39-402e-b0a1-d5ca5fc4e630. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 732.831327] env[61906]: DEBUG oslo_concurrency.lockutils [req-5fe76ab6-68ee-4e4a-8fdd-8a3fe18b5125 req-0219be2c-3457-4b31-9283-f68592a323d9 service nova] Acquiring lock "refresh_cache-c15bc75b-87dc-4602-81ea-dbebce1b7763" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.831469] env[61906]: DEBUG oslo_concurrency.lockutils [req-5fe76ab6-68ee-4e4a-8fdd-8a3fe18b5125 req-0219be2c-3457-4b31-9283-f68592a323d9 service nova] Acquired lock "refresh_cache-c15bc75b-87dc-4602-81ea-dbebce1b7763" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.831627] env[61906]: DEBUG nova.network.neutron [req-5fe76ab6-68ee-4e4a-8fdd-8a3fe18b5125 req-0219be2c-3457-4b31-9283-f68592a323d9 service nova] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Refreshing network info cache for port 8aa8e05d-7e39-402e-b0a1-d5ca5fc4e630 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 732.948445] env[61906]: ERROR nova.compute.manager [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8aa8e05d-7e39-402e-b0a1-d5ca5fc4e630, please check neutron logs for more information. [ 732.948445] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 732.948445] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.948445] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 732.948445] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.948445] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 732.948445] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.948445] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 732.948445] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.948445] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 732.948445] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.948445] env[61906]: ERROR nova.compute.manager raise self.value [ 732.948445] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.948445] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 732.948445] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.948445] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 732.948955] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.948955] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 732.948955] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8aa8e05d-7e39-402e-b0a1-d5ca5fc4e630, please check neutron logs for more information. [ 732.948955] env[61906]: ERROR nova.compute.manager [ 732.948955] env[61906]: Traceback (most recent call last): [ 732.948955] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 732.948955] env[61906]: listener.cb(fileno) [ 732.948955] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.948955] env[61906]: result = function(*args, **kwargs) [ 732.948955] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 732.948955] env[61906]: return func(*args, **kwargs) [ 732.948955] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.948955] env[61906]: raise e [ 732.948955] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.948955] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 732.948955] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.948955] env[61906]: created_port_ids = self._update_ports_for_instance( [ 732.948955] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.948955] env[61906]: with excutils.save_and_reraise_exception(): [ 732.948955] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.948955] env[61906]: self.force_reraise() [ 732.948955] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.948955] env[61906]: raise self.value [ 732.948955] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.948955] env[61906]: updated_port = self._update_port( [ 732.948955] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.948955] env[61906]: _ensure_no_port_binding_failure(port) [ 732.948955] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.948955] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 732.949823] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 8aa8e05d-7e39-402e-b0a1-d5ca5fc4e630, please check neutron logs for more information. [ 732.949823] env[61906]: Removing descriptor: 16 [ 732.949823] env[61906]: ERROR nova.compute.manager [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8aa8e05d-7e39-402e-b0a1-d5ca5fc4e630, please check neutron logs for more information. [ 732.949823] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Traceback (most recent call last): [ 732.949823] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 732.949823] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] yield resources [ 732.949823] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 732.949823] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] self.driver.spawn(context, instance, image_meta, [ 732.949823] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 732.949823] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] self._vmops.spawn(context, instance, image_meta, injected_files, [ 732.949823] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 732.949823] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] vm_ref = self.build_virtual_machine(instance, [ 732.950218] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 732.950218] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] vif_infos = vmwarevif.get_vif_info(self._session, [ 732.950218] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 732.950218] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] for vif in network_info: [ 732.950218] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 732.950218] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] return self._sync_wrapper(fn, *args, **kwargs) [ 732.950218] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 732.950218] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] self.wait() [ 732.950218] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 732.950218] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] self[:] = self._gt.wait() [ 732.950218] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 732.950218] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] return self._exit_event.wait() [ 732.950218] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 732.950611] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] result = hub.switch() [ 732.950611] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 732.950611] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] return self.greenlet.switch() [ 732.950611] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.950611] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] result = function(*args, **kwargs) [ 732.950611] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 732.950611] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] return func(*args, **kwargs) [ 732.950611] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.950611] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] raise e [ 732.950611] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.950611] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] nwinfo = self.network_api.allocate_for_instance( [ 732.950611] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.950611] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] created_port_ids = self._update_ports_for_instance( [ 732.951037] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.951037] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] with excutils.save_and_reraise_exception(): [ 732.951037] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.951037] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] self.force_reraise() [ 732.951037] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.951037] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] raise self.value [ 732.951037] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.951037] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] updated_port = self._update_port( [ 732.951037] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.951037] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] _ensure_no_port_binding_failure(port) [ 732.951037] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.951037] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] raise exception.PortBindingFailed(port_id=port['id']) [ 732.951399] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] nova.exception.PortBindingFailed: Binding failed for port 8aa8e05d-7e39-402e-b0a1-d5ca5fc4e630, please check neutron logs for more information. [ 732.951399] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] [ 732.951399] env[61906]: INFO nova.compute.manager [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Terminating instance [ 732.952312] env[61906]: DEBUG oslo_concurrency.lockutils [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Acquiring lock "refresh_cache-c15bc75b-87dc-4602-81ea-dbebce1b7763" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.002015] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.140163] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.486s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.141125] env[61906]: DEBUG nova.compute.manager [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 733.143703] env[61906]: DEBUG oslo_concurrency.lockutils [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.479s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.145082] env[61906]: INFO nova.compute.claims [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 733.354233] env[61906]: DEBUG nova.network.neutron [req-5fe76ab6-68ee-4e4a-8fdd-8a3fe18b5125 req-0219be2c-3457-4b31-9283-f68592a323d9 service nova] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.448139] env[61906]: DEBUG nova.network.neutron [req-5fe76ab6-68ee-4e4a-8fdd-8a3fe18b5125 req-0219be2c-3457-4b31-9283-f68592a323d9 service nova] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.649060] env[61906]: DEBUG nova.compute.utils [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 733.650852] env[61906]: DEBUG nova.compute.manager [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 733.650852] env[61906]: DEBUG nova.network.neutron [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 733.708527] env[61906]: DEBUG nova.policy [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1b889ef55dbc4f4d9d88c0d57a8eaaaf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0983a7da8870442b86b96cfde74bf6a0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 733.933034] env[61906]: DEBUG oslo_concurrency.lockutils [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "e0afe101-f4a4-4dc8-a666-9a7793dbf8a6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.933219] env[61906]: DEBUG oslo_concurrency.lockutils [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "e0afe101-f4a4-4dc8-a666-9a7793dbf8a6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.950753] env[61906]: DEBUG oslo_concurrency.lockutils [req-5fe76ab6-68ee-4e4a-8fdd-8a3fe18b5125 req-0219be2c-3457-4b31-9283-f68592a323d9 service nova] Releasing lock "refresh_cache-c15bc75b-87dc-4602-81ea-dbebce1b7763" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.951195] env[61906]: DEBUG oslo_concurrency.lockutils [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Acquired lock "refresh_cache-c15bc75b-87dc-4602-81ea-dbebce1b7763" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.951365] env[61906]: DEBUG nova.network.neutron [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 734.006234] env[61906]: DEBUG nova.network.neutron [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Successfully created port: 766bdfb7-99ba-4eea-a90f-0da913c62e82 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 734.155936] env[61906]: DEBUG nova.compute.manager [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 734.469912] env[61906]: DEBUG nova.network.neutron [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 734.534100] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df615bf1-5923-4929-9360-030675a476ae {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.541298] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f986cbf-e1c6-40e4-9dcc-76ec33a37aad {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.572648] env[61906]: DEBUG nova.network.neutron [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.574328] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b2d1e5-83c5-4f90-99ff-3a56f531670d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.582533] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abdc84aa-47e0-49fa-b788-c2c6fec03552 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.596888] env[61906]: DEBUG nova.compute.provider_tree [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.877133] env[61906]: DEBUG nova.compute.manager [req-603bde30-b90b-4d8d-99f8-b8d0978a6d72 req-549f9dbd-937f-4465-a47c-bed6815c413d service nova] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Received event network-vif-deleted-8aa8e05d-7e39-402e-b0a1-d5ca5fc4e630 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 734.877133] env[61906]: DEBUG nova.compute.manager [req-603bde30-b90b-4d8d-99f8-b8d0978a6d72 req-549f9dbd-937f-4465-a47c-bed6815c413d service nova] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Received event network-changed-766bdfb7-99ba-4eea-a90f-0da913c62e82 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 734.877133] env[61906]: DEBUG nova.compute.manager [req-603bde30-b90b-4d8d-99f8-b8d0978a6d72 req-549f9dbd-937f-4465-a47c-bed6815c413d service nova] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Refreshing instance network info cache due to event network-changed-766bdfb7-99ba-4eea-a90f-0da913c62e82. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 734.877538] env[61906]: DEBUG oslo_concurrency.lockutils [req-603bde30-b90b-4d8d-99f8-b8d0978a6d72 req-549f9dbd-937f-4465-a47c-bed6815c413d service nova] Acquiring lock "refresh_cache-c9cb7335-8977-4ddf-9d0e-ff2166abf8e7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.877719] env[61906]: DEBUG oslo_concurrency.lockutils [req-603bde30-b90b-4d8d-99f8-b8d0978a6d72 req-549f9dbd-937f-4465-a47c-bed6815c413d service nova] Acquired lock "refresh_cache-c9cb7335-8977-4ddf-9d0e-ff2166abf8e7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.877922] env[61906]: DEBUG nova.network.neutron [req-603bde30-b90b-4d8d-99f8-b8d0978a6d72 req-549f9dbd-937f-4465-a47c-bed6815c413d service nova] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Refreshing network info cache for port 766bdfb7-99ba-4eea-a90f-0da913c62e82 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 734.977329] env[61906]: ERROR nova.compute.manager [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 766bdfb7-99ba-4eea-a90f-0da913c62e82, please check neutron logs for more information. [ 734.977329] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 734.977329] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.977329] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 734.977329] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.977329] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 734.977329] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.977329] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 734.977329] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.977329] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 734.977329] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.977329] env[61906]: ERROR nova.compute.manager raise self.value [ 734.977329] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.977329] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 734.977329] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.977329] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 734.977941] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.977941] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 734.977941] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 766bdfb7-99ba-4eea-a90f-0da913c62e82, please check neutron logs for more information. [ 734.977941] env[61906]: ERROR nova.compute.manager [ 734.977941] env[61906]: Traceback (most recent call last): [ 734.977941] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 734.977941] env[61906]: listener.cb(fileno) [ 734.977941] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 734.977941] env[61906]: result = function(*args, **kwargs) [ 734.977941] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 734.977941] env[61906]: return func(*args, **kwargs) [ 734.977941] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 734.977941] env[61906]: raise e [ 734.977941] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.977941] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 734.977941] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.977941] env[61906]: created_port_ids = self._update_ports_for_instance( [ 734.977941] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.977941] env[61906]: with excutils.save_and_reraise_exception(): [ 734.977941] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.977941] env[61906]: self.force_reraise() [ 734.977941] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.977941] env[61906]: raise self.value [ 734.977941] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.977941] env[61906]: updated_port = self._update_port( [ 734.977941] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.977941] env[61906]: _ensure_no_port_binding_failure(port) [ 734.977941] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.977941] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 734.978662] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 766bdfb7-99ba-4eea-a90f-0da913c62e82, please check neutron logs for more information. [ 734.978662] env[61906]: Removing descriptor: 16 [ 735.079053] env[61906]: DEBUG oslo_concurrency.lockutils [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Releasing lock "refresh_cache-c15bc75b-87dc-4602-81ea-dbebce1b7763" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.079183] env[61906]: DEBUG nova.compute.manager [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 735.079330] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 735.079641] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c64a331f-9a2a-4387-946a-3ffb77c1539a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.088412] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed514e53-99e5-49cb-9efd-0261af1246d2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.100546] env[61906]: DEBUG nova.scheduler.client.report [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 735.114386] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c15bc75b-87dc-4602-81ea-dbebce1b7763 could not be found. [ 735.114591] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 735.114766] env[61906]: INFO nova.compute.manager [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Took 0.04 seconds to destroy the instance on the hypervisor. [ 735.115009] env[61906]: DEBUG oslo.service.loopingcall [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 735.115759] env[61906]: DEBUG nova.compute.manager [-] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 735.115861] env[61906]: DEBUG nova.network.neutron [-] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 735.135940] env[61906]: DEBUG nova.network.neutron [-] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.168320] env[61906]: DEBUG nova.compute.manager [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 735.192918] env[61906]: DEBUG nova.virt.hardware [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 735.193187] env[61906]: DEBUG nova.virt.hardware [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 735.193340] env[61906]: DEBUG nova.virt.hardware [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 735.193521] env[61906]: DEBUG nova.virt.hardware [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 735.193665] env[61906]: DEBUG nova.virt.hardware [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 735.193809] env[61906]: DEBUG nova.virt.hardware [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 735.194015] env[61906]: DEBUG nova.virt.hardware [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 735.194183] env[61906]: DEBUG nova.virt.hardware [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 735.194349] env[61906]: DEBUG nova.virt.hardware [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 735.194508] env[61906]: DEBUG nova.virt.hardware [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 735.194676] env[61906]: DEBUG nova.virt.hardware [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 735.195523] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d8c0dc4-5bab-425a-af07-da9bbaadab56 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.203532] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49288651-1517-481d-a106-9523b68ce7e8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.216915] env[61906]: ERROR nova.compute.manager [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 766bdfb7-99ba-4eea-a90f-0da913c62e82, please check neutron logs for more information. [ 735.216915] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Traceback (most recent call last): [ 735.216915] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 735.216915] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] yield resources [ 735.216915] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 735.216915] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] self.driver.spawn(context, instance, image_meta, [ 735.216915] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 735.216915] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 735.216915] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 735.216915] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] vm_ref = self.build_virtual_machine(instance, [ 735.216915] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 735.217275] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] vif_infos = vmwarevif.get_vif_info(self._session, [ 735.217275] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 735.217275] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] for vif in network_info: [ 735.217275] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 735.217275] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] return self._sync_wrapper(fn, *args, **kwargs) [ 735.217275] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 735.217275] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] self.wait() [ 735.217275] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 735.217275] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] self[:] = self._gt.wait() [ 735.217275] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 735.217275] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] return self._exit_event.wait() [ 735.217275] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 735.217275] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] current.throw(*self._exc) [ 735.217705] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.217705] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] result = function(*args, **kwargs) [ 735.217705] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 735.217705] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] return func(*args, **kwargs) [ 735.217705] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 735.217705] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] raise e [ 735.217705] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.217705] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] nwinfo = self.network_api.allocate_for_instance( [ 735.217705] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 735.217705] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] created_port_ids = self._update_ports_for_instance( [ 735.217705] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 735.217705] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] with excutils.save_and_reraise_exception(): [ 735.217705] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.218125] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] self.force_reraise() [ 735.218125] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.218125] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] raise self.value [ 735.218125] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 735.218125] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] updated_port = self._update_port( [ 735.218125] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.218125] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] _ensure_no_port_binding_failure(port) [ 735.218125] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.218125] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] raise exception.PortBindingFailed(port_id=port['id']) [ 735.218125] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] nova.exception.PortBindingFailed: Binding failed for port 766bdfb7-99ba-4eea-a90f-0da913c62e82, please check neutron logs for more information. [ 735.218125] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] [ 735.218125] env[61906]: INFO nova.compute.manager [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Terminating instance [ 735.219471] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Acquiring lock "refresh_cache-c9cb7335-8977-4ddf-9d0e-ff2166abf8e7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.399160] env[61906]: DEBUG nova.network.neutron [req-603bde30-b90b-4d8d-99f8-b8d0978a6d72 req-549f9dbd-937f-4465-a47c-bed6815c413d service nova] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.492915] env[61906]: DEBUG nova.network.neutron [req-603bde30-b90b-4d8d-99f8-b8d0978a6d72 req-549f9dbd-937f-4465-a47c-bed6815c413d service nova] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.605707] env[61906]: DEBUG oslo_concurrency.lockutils [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.462s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.606080] env[61906]: DEBUG nova.compute.manager [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 735.609754] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.617s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.638500] env[61906]: DEBUG nova.network.neutron [-] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.996178] env[61906]: DEBUG oslo_concurrency.lockutils [req-603bde30-b90b-4d8d-99f8-b8d0978a6d72 req-549f9dbd-937f-4465-a47c-bed6815c413d service nova] Releasing lock "refresh_cache-c9cb7335-8977-4ddf-9d0e-ff2166abf8e7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.996628] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Acquired lock "refresh_cache-c9cb7335-8977-4ddf-9d0e-ff2166abf8e7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.996817] env[61906]: DEBUG nova.network.neutron [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 736.114949] env[61906]: DEBUG nova.compute.utils [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 736.119712] env[61906]: DEBUG nova.compute.manager [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 736.119712] env[61906]: DEBUG nova.network.neutron [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 736.142867] env[61906]: INFO nova.compute.manager [-] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Took 1.03 seconds to deallocate network for instance. [ 736.145700] env[61906]: DEBUG nova.compute.claims [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 736.145872] env[61906]: DEBUG oslo_concurrency.lockutils [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.159700] env[61906]: DEBUG nova.policy [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f33ae649c28a45688c8662d4b3a29c9a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ac8ee7432c4a414d9d6d7ce5f60ced7a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 736.437604] env[61906]: DEBUG nova.network.neutron [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Successfully created port: f92e1b5b-fa31-476c-84aa-eade9006bbd2 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 736.449150] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c260a269-4fee-47bb-a595-bcca910b851f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.458345] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-903a0648-ed2c-4e37-b5d6-92a71292b36c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.492675] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a75476c9-06c1-4d5e-b4f0-1a22b008a77a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.502998] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3134e334-fda5-42d5-8039-f7f0f76a59e6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.515747] env[61906]: DEBUG nova.compute.provider_tree [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 736.621691] env[61906]: DEBUG nova.compute.manager [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 736.679337] env[61906]: DEBUG nova.network.neutron [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.762993] env[61906]: DEBUG nova.network.neutron [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.910409] env[61906]: DEBUG nova.compute.manager [req-1d13eacc-3db4-4493-a01b-d1f4146dd473 req-35deb469-1fe2-4a3b-b1f7-88141a4b57e4 service nova] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Received event network-vif-deleted-766bdfb7-99ba-4eea-a90f-0da913c62e82 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 737.019592] env[61906]: DEBUG nova.scheduler.client.report [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 737.266023] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Releasing lock "refresh_cache-c9cb7335-8977-4ddf-9d0e-ff2166abf8e7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.266480] env[61906]: DEBUG nova.compute.manager [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 737.266672] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 737.266975] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-92476180-a48c-49ee-844b-f8bbe8cea402 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.276929] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81c29942-57a1-4a48-a4e8-0470219aae1d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.303314] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c9cb7335-8977-4ddf-9d0e-ff2166abf8e7 could not be found. [ 737.303560] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 737.303736] env[61906]: INFO nova.compute.manager [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 737.303976] env[61906]: DEBUG oslo.service.loopingcall [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 737.304215] env[61906]: DEBUG nova.compute.manager [-] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 737.304305] env[61906]: DEBUG nova.network.neutron [-] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 737.326441] env[61906]: DEBUG nova.network.neutron [-] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.436445] env[61906]: ERROR nova.compute.manager [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f92e1b5b-fa31-476c-84aa-eade9006bbd2, please check neutron logs for more information. [ 737.436445] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 737.436445] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 737.436445] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 737.436445] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 737.436445] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 737.436445] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 737.436445] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 737.436445] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.436445] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 737.436445] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.436445] env[61906]: ERROR nova.compute.manager raise self.value [ 737.436445] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 737.436445] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 737.436445] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.436445] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 737.436945] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.436945] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 737.436945] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f92e1b5b-fa31-476c-84aa-eade9006bbd2, please check neutron logs for more information. [ 737.436945] env[61906]: ERROR nova.compute.manager [ 737.436945] env[61906]: Traceback (most recent call last): [ 737.436945] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 737.436945] env[61906]: listener.cb(fileno) [ 737.436945] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 737.436945] env[61906]: result = function(*args, **kwargs) [ 737.436945] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 737.436945] env[61906]: return func(*args, **kwargs) [ 737.436945] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 737.436945] env[61906]: raise e [ 737.436945] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 737.436945] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 737.436945] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 737.436945] env[61906]: created_port_ids = self._update_ports_for_instance( [ 737.436945] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 737.436945] env[61906]: with excutils.save_and_reraise_exception(): [ 737.436945] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.436945] env[61906]: self.force_reraise() [ 737.436945] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.436945] env[61906]: raise self.value [ 737.436945] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 737.436945] env[61906]: updated_port = self._update_port( [ 737.436945] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.436945] env[61906]: _ensure_no_port_binding_failure(port) [ 737.436945] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.436945] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 737.437817] env[61906]: nova.exception.PortBindingFailed: Binding failed for port f92e1b5b-fa31-476c-84aa-eade9006bbd2, please check neutron logs for more information. [ 737.437817] env[61906]: Removing descriptor: 16 [ 737.528359] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.918s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.529023] env[61906]: ERROR nova.compute.manager [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 774f613b-ada5-476b-95e0-cbb8bb0f4da4, please check neutron logs for more information. [ 737.529023] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Traceback (most recent call last): [ 737.529023] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 737.529023] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] self.driver.spawn(context, instance, image_meta, [ 737.529023] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 737.529023] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] self._vmops.spawn(context, instance, image_meta, injected_files, [ 737.529023] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 737.529023] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] vm_ref = self.build_virtual_machine(instance, [ 737.529023] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 737.529023] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] vif_infos = vmwarevif.get_vif_info(self._session, [ 737.529023] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 737.529337] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] for vif in network_info: [ 737.529337] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 737.529337] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] return self._sync_wrapper(fn, *args, **kwargs) [ 737.529337] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 737.529337] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] self.wait() [ 737.529337] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 737.529337] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] self[:] = self._gt.wait() [ 737.529337] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 737.529337] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] return self._exit_event.wait() [ 737.529337] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 737.529337] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] result = hub.switch() [ 737.529337] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 737.529337] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] return self.greenlet.switch() [ 737.529655] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 737.529655] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] result = function(*args, **kwargs) [ 737.529655] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 737.529655] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] return func(*args, **kwargs) [ 737.529655] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 737.529655] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] raise e [ 737.529655] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 737.529655] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] nwinfo = self.network_api.allocate_for_instance( [ 737.529655] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 737.529655] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] created_port_ids = self._update_ports_for_instance( [ 737.529655] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 737.529655] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] with excutils.save_and_reraise_exception(): [ 737.529655] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.529985] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] self.force_reraise() [ 737.529985] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.529985] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] raise self.value [ 737.529985] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 737.529985] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] updated_port = self._update_port( [ 737.529985] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.529985] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] _ensure_no_port_binding_failure(port) [ 737.529985] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.529985] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] raise exception.PortBindingFailed(port_id=port['id']) [ 737.529985] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] nova.exception.PortBindingFailed: Binding failed for port 774f613b-ada5-476b-95e0-cbb8bb0f4da4, please check neutron logs for more information. [ 737.529985] env[61906]: ERROR nova.compute.manager [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] [ 737.530451] env[61906]: DEBUG nova.compute.utils [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Binding failed for port 774f613b-ada5-476b-95e0-cbb8bb0f4da4, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 737.531159] env[61906]: DEBUG oslo_concurrency.lockutils [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.110s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.535248] env[61906]: DEBUG nova.compute.manager [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Build of instance fe14efda-3b92-4045-8eef-0e5e92a9c538 was re-scheduled: Binding failed for port 774f613b-ada5-476b-95e0-cbb8bb0f4da4, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 737.535711] env[61906]: DEBUG nova.compute.manager [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 737.536240] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Acquiring lock "refresh_cache-fe14efda-3b92-4045-8eef-0e5e92a9c538" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.536240] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Acquired lock "refresh_cache-fe14efda-3b92-4045-8eef-0e5e92a9c538" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.536240] env[61906]: DEBUG nova.network.neutron [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 737.632361] env[61906]: DEBUG nova.compute.manager [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 737.657904] env[61906]: DEBUG nova.virt.hardware [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 737.658165] env[61906]: DEBUG nova.virt.hardware [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 737.658324] env[61906]: DEBUG nova.virt.hardware [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 737.658503] env[61906]: DEBUG nova.virt.hardware [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 737.658700] env[61906]: DEBUG nova.virt.hardware [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 737.658934] env[61906]: DEBUG nova.virt.hardware [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 737.659192] env[61906]: DEBUG nova.virt.hardware [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 737.659357] env[61906]: DEBUG nova.virt.hardware [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 737.659525] env[61906]: DEBUG nova.virt.hardware [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 737.659684] env[61906]: DEBUG nova.virt.hardware [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 737.659855] env[61906]: DEBUG nova.virt.hardware [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 737.660718] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d192319-9fde-4c9c-814a-e806a4303b0d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.668786] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3107f871-cae6-42c7-b542-cc0da4799d7e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.682313] env[61906]: ERROR nova.compute.manager [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f92e1b5b-fa31-476c-84aa-eade9006bbd2, please check neutron logs for more information. [ 737.682313] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Traceback (most recent call last): [ 737.682313] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 737.682313] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] yield resources [ 737.682313] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 737.682313] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] self.driver.spawn(context, instance, image_meta, [ 737.682313] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 737.682313] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] self._vmops.spawn(context, instance, image_meta, injected_files, [ 737.682313] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 737.682313] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] vm_ref = self.build_virtual_machine(instance, [ 737.682313] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 737.682704] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] vif_infos = vmwarevif.get_vif_info(self._session, [ 737.682704] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 737.682704] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] for vif in network_info: [ 737.682704] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 737.682704] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] return self._sync_wrapper(fn, *args, **kwargs) [ 737.682704] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 737.682704] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] self.wait() [ 737.682704] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 737.682704] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] self[:] = self._gt.wait() [ 737.682704] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 737.682704] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] return self._exit_event.wait() [ 737.682704] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 737.682704] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] current.throw(*self._exc) [ 737.683110] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 737.683110] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] result = function(*args, **kwargs) [ 737.683110] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 737.683110] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] return func(*args, **kwargs) [ 737.683110] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 737.683110] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] raise e [ 737.683110] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 737.683110] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] nwinfo = self.network_api.allocate_for_instance( [ 737.683110] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 737.683110] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] created_port_ids = self._update_ports_for_instance( [ 737.683110] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 737.683110] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] with excutils.save_and_reraise_exception(): [ 737.683110] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.683503] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] self.force_reraise() [ 737.683503] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.683503] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] raise self.value [ 737.683503] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 737.683503] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] updated_port = self._update_port( [ 737.683503] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.683503] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] _ensure_no_port_binding_failure(port) [ 737.683503] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.683503] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] raise exception.PortBindingFailed(port_id=port['id']) [ 737.683503] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] nova.exception.PortBindingFailed: Binding failed for port f92e1b5b-fa31-476c-84aa-eade9006bbd2, please check neutron logs for more information. [ 737.683503] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] [ 737.683503] env[61906]: INFO nova.compute.manager [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Terminating instance [ 737.684642] env[61906]: DEBUG oslo_concurrency.lockutils [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Acquiring lock "refresh_cache-c4b41c42-e01f-464a-9423-1c0fcca75586" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.684802] env[61906]: DEBUG oslo_concurrency.lockutils [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Acquired lock "refresh_cache-c4b41c42-e01f-464a-9423-1c0fcca75586" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.684968] env[61906]: DEBUG nova.network.neutron [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 737.830342] env[61906]: DEBUG nova.network.neutron [-] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.055268] env[61906]: DEBUG nova.network.neutron [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.134432] env[61906]: DEBUG nova.network.neutron [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.203529] env[61906]: DEBUG nova.network.neutron [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.286855] env[61906]: DEBUG nova.network.neutron [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.332462] env[61906]: INFO nova.compute.manager [-] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Took 1.03 seconds to deallocate network for instance. [ 738.334873] env[61906]: DEBUG nova.compute.claims [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 738.335069] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.356264] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b62bb7c-4a9d-46b2-b473-69d613951d0b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.363748] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a0a5185-21bd-4697-a22d-d0da029e4fa2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.403367] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61e6c355-a18f-4f77-a0db-10475fefb3c4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.411462] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600f52e9-d17d-4df7-b1be-71047bc34fdd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.426593] env[61906]: DEBUG nova.compute.provider_tree [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 738.638742] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Releasing lock "refresh_cache-fe14efda-3b92-4045-8eef-0e5e92a9c538" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.639092] env[61906]: DEBUG nova.compute.manager [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 738.639277] env[61906]: DEBUG nova.compute.manager [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 738.639447] env[61906]: DEBUG nova.network.neutron [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 738.652090] env[61906]: DEBUG nova.network.neutron [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.791891] env[61906]: DEBUG oslo_concurrency.lockutils [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Releasing lock "refresh_cache-c4b41c42-e01f-464a-9423-1c0fcca75586" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.792409] env[61906]: DEBUG nova.compute.manager [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 738.792619] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 738.792918] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e61f4b48-7112-4720-86a1-f254ecfaa1a9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.802298] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-decf0b2c-68d6-43d3-9fc1-a060464deeb4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.822990] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c4b41c42-e01f-464a-9423-1c0fcca75586 could not be found. [ 738.823234] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 738.823413] env[61906]: INFO nova.compute.manager [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Took 0.03 seconds to destroy the instance on the hypervisor. [ 738.823656] env[61906]: DEBUG oslo.service.loopingcall [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 738.823878] env[61906]: DEBUG nova.compute.manager [-] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 738.823978] env[61906]: DEBUG nova.network.neutron [-] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 738.839670] env[61906]: DEBUG nova.network.neutron [-] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.929422] env[61906]: DEBUG nova.scheduler.client.report [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 738.963815] env[61906]: DEBUG nova.compute.manager [req-033f342e-7624-4963-ae4b-b1b92241e085 req-52436931-adc0-4851-b316-a0152d55ed11 service nova] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Received event network-changed-f92e1b5b-fa31-476c-84aa-eade9006bbd2 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 738.964068] env[61906]: DEBUG nova.compute.manager [req-033f342e-7624-4963-ae4b-b1b92241e085 req-52436931-adc0-4851-b316-a0152d55ed11 service nova] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Refreshing instance network info cache due to event network-changed-f92e1b5b-fa31-476c-84aa-eade9006bbd2. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 738.964346] env[61906]: DEBUG oslo_concurrency.lockutils [req-033f342e-7624-4963-ae4b-b1b92241e085 req-52436931-adc0-4851-b316-a0152d55ed11 service nova] Acquiring lock "refresh_cache-c4b41c42-e01f-464a-9423-1c0fcca75586" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.964524] env[61906]: DEBUG oslo_concurrency.lockutils [req-033f342e-7624-4963-ae4b-b1b92241e085 req-52436931-adc0-4851-b316-a0152d55ed11 service nova] Acquired lock "refresh_cache-c4b41c42-e01f-464a-9423-1c0fcca75586" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.964718] env[61906]: DEBUG nova.network.neutron [req-033f342e-7624-4963-ae4b-b1b92241e085 req-52436931-adc0-4851-b316-a0152d55ed11 service nova] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Refreshing network info cache for port f92e1b5b-fa31-476c-84aa-eade9006bbd2 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 739.154797] env[61906]: DEBUG nova.network.neutron [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.342555] env[61906]: DEBUG nova.network.neutron [-] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.434698] env[61906]: DEBUG oslo_concurrency.lockutils [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.904s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.436023] env[61906]: ERROR nova.compute.manager [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ebd40d4d-5c50-4ebd-81a4-44ec4d9bfb2e, please check neutron logs for more information. [ 739.436023] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Traceback (most recent call last): [ 739.436023] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 739.436023] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] self.driver.spawn(context, instance, image_meta, [ 739.436023] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 739.436023] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 739.436023] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 739.436023] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] vm_ref = self.build_virtual_machine(instance, [ 739.436023] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 739.436023] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] vif_infos = vmwarevif.get_vif_info(self._session, [ 739.436023] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 739.436442] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] for vif in network_info: [ 739.436442] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 739.436442] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] return self._sync_wrapper(fn, *args, **kwargs) [ 739.436442] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 739.436442] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] self.wait() [ 739.436442] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 739.436442] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] self[:] = self._gt.wait() [ 739.436442] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 739.436442] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] return self._exit_event.wait() [ 739.436442] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 739.436442] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] result = hub.switch() [ 739.436442] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 739.436442] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] return self.greenlet.switch() [ 739.436841] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.436841] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] result = function(*args, **kwargs) [ 739.436841] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 739.436841] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] return func(*args, **kwargs) [ 739.436841] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 739.436841] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] raise e [ 739.436841] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.436841] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] nwinfo = self.network_api.allocate_for_instance( [ 739.436841] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.436841] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] created_port_ids = self._update_ports_for_instance( [ 739.436841] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.436841] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] with excutils.save_and_reraise_exception(): [ 739.436841] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.437227] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] self.force_reraise() [ 739.437227] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.437227] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] raise self.value [ 739.437227] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.437227] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] updated_port = self._update_port( [ 739.437227] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.437227] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] _ensure_no_port_binding_failure(port) [ 739.437227] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.437227] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] raise exception.PortBindingFailed(port_id=port['id']) [ 739.437227] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] nova.exception.PortBindingFailed: Binding failed for port ebd40d4d-5c50-4ebd-81a4-44ec4d9bfb2e, please check neutron logs for more information. [ 739.437227] env[61906]: ERROR nova.compute.manager [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] [ 739.437547] env[61906]: DEBUG nova.compute.utils [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Binding failed for port ebd40d4d-5c50-4ebd-81a4-44ec4d9bfb2e, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 739.437547] env[61906]: DEBUG oslo_concurrency.lockutils [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.604s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.440338] env[61906]: DEBUG nova.compute.manager [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Build of instance 9e665b58-69e7-4ab9-b109-ab27725e66cc was re-scheduled: Binding failed for port ebd40d4d-5c50-4ebd-81a4-44ec4d9bfb2e, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 739.440460] env[61906]: DEBUG nova.compute.manager [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 739.440672] env[61906]: DEBUG oslo_concurrency.lockutils [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Acquiring lock "refresh_cache-9e665b58-69e7-4ab9-b109-ab27725e66cc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.440818] env[61906]: DEBUG oslo_concurrency.lockutils [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Acquired lock "refresh_cache-9e665b58-69e7-4ab9-b109-ab27725e66cc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.441030] env[61906]: DEBUG nova.network.neutron [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 739.484068] env[61906]: DEBUG nova.network.neutron [req-033f342e-7624-4963-ae4b-b1b92241e085 req-52436931-adc0-4851-b316-a0152d55ed11 service nova] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.577822] env[61906]: DEBUG nova.network.neutron [req-033f342e-7624-4963-ae4b-b1b92241e085 req-52436931-adc0-4851-b316-a0152d55ed11 service nova] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.657862] env[61906]: INFO nova.compute.manager [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] [instance: fe14efda-3b92-4045-8eef-0e5e92a9c538] Took 1.02 seconds to deallocate network for instance. [ 739.845350] env[61906]: INFO nova.compute.manager [-] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Took 1.02 seconds to deallocate network for instance. [ 739.847786] env[61906]: DEBUG nova.compute.claims [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 739.847980] env[61906]: DEBUG oslo_concurrency.lockutils [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.963477] env[61906]: DEBUG nova.network.neutron [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.079445] env[61906]: DEBUG nova.network.neutron [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.081740] env[61906]: DEBUG oslo_concurrency.lockutils [req-033f342e-7624-4963-ae4b-b1b92241e085 req-52436931-adc0-4851-b316-a0152d55ed11 service nova] Releasing lock "refresh_cache-c4b41c42-e01f-464a-9423-1c0fcca75586" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.082089] env[61906]: DEBUG nova.compute.manager [req-033f342e-7624-4963-ae4b-b1b92241e085 req-52436931-adc0-4851-b316-a0152d55ed11 service nova] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Received event network-vif-deleted-f92e1b5b-fa31-476c-84aa-eade9006bbd2 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 740.278447] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4be0129-ac7a-49a6-872a-d7ba7c101032 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.286352] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a480a6-9fe3-47a8-a305-b9afa764b6b1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.315612] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a13b249a-5684-474f-8d9f-bf1c457e071e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.324040] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59c2e68d-ad10-41a1-b22b-90fe366d182d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.339015] env[61906]: DEBUG nova.compute.provider_tree [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 740.583284] env[61906]: DEBUG oslo_concurrency.lockutils [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Releasing lock "refresh_cache-9e665b58-69e7-4ab9-b109-ab27725e66cc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.583549] env[61906]: DEBUG nova.compute.manager [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 740.583738] env[61906]: DEBUG nova.compute.manager [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 740.583913] env[61906]: DEBUG nova.network.neutron [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 740.598619] env[61906]: DEBUG nova.network.neutron [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.689637] env[61906]: INFO nova.scheduler.client.report [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Deleted allocations for instance fe14efda-3b92-4045-8eef-0e5e92a9c538 [ 740.842233] env[61906]: DEBUG nova.scheduler.client.report [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 741.101112] env[61906]: DEBUG nova.network.neutron [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.200193] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5c503e7f-d924-4cf1-bd39-48bef3e1c7ae tempest-ServerPasswordTestJSON-1642087198 tempest-ServerPasswordTestJSON-1642087198-project-member] Lock "fe14efda-3b92-4045-8eef-0e5e92a9c538" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 146.102s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.348072] env[61906]: DEBUG oslo_concurrency.lockutils [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.910s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.348072] env[61906]: ERROR nova.compute.manager [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0ccf6630-3159-48b7-8f73-0112b29176eb, please check neutron logs for more information. [ 741.348072] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Traceback (most recent call last): [ 741.348072] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 741.348072] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] self.driver.spawn(context, instance, image_meta, [ 741.348072] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 741.348072] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] self._vmops.spawn(context, instance, image_meta, injected_files, [ 741.348072] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 741.348072] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] vm_ref = self.build_virtual_machine(instance, [ 741.348690] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 741.348690] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] vif_infos = vmwarevif.get_vif_info(self._session, [ 741.348690] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 741.348690] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] for vif in network_info: [ 741.348690] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 741.348690] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] return self._sync_wrapper(fn, *args, **kwargs) [ 741.348690] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 741.348690] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] self.wait() [ 741.348690] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 741.348690] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] self[:] = self._gt.wait() [ 741.348690] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 741.348690] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] return self._exit_event.wait() [ 741.348690] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 741.349040] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] result = hub.switch() [ 741.349040] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 741.349040] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] return self.greenlet.switch() [ 741.349040] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 741.349040] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] result = function(*args, **kwargs) [ 741.349040] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 741.349040] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] return func(*args, **kwargs) [ 741.349040] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 741.349040] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] raise e [ 741.349040] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 741.349040] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] nwinfo = self.network_api.allocate_for_instance( [ 741.349040] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 741.349040] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] created_port_ids = self._update_ports_for_instance( [ 741.349566] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 741.349566] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] with excutils.save_and_reraise_exception(): [ 741.349566] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 741.349566] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] self.force_reraise() [ 741.349566] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 741.349566] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] raise self.value [ 741.349566] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 741.349566] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] updated_port = self._update_port( [ 741.349566] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 741.349566] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] _ensure_no_port_binding_failure(port) [ 741.349566] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 741.349566] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] raise exception.PortBindingFailed(port_id=port['id']) [ 741.349922] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] nova.exception.PortBindingFailed: Binding failed for port 0ccf6630-3159-48b7-8f73-0112b29176eb, please check neutron logs for more information. [ 741.349922] env[61906]: ERROR nova.compute.manager [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] [ 741.349922] env[61906]: DEBUG nova.compute.utils [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Binding failed for port 0ccf6630-3159-48b7-8f73-0112b29176eb, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 741.350035] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.147s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.351515] env[61906]: INFO nova.compute.claims [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 741.354257] env[61906]: DEBUG nova.compute.manager [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Build of instance 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27 was re-scheduled: Binding failed for port 0ccf6630-3159-48b7-8f73-0112b29176eb, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 741.354691] env[61906]: DEBUG nova.compute.manager [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 741.354915] env[61906]: DEBUG oslo_concurrency.lockutils [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Acquiring lock "refresh_cache-8d2459c0-279c-4ea4-8cbc-6c69cfa35b27" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.355093] env[61906]: DEBUG oslo_concurrency.lockutils [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Acquired lock "refresh_cache-8d2459c0-279c-4ea4-8cbc-6c69cfa35b27" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.355257] env[61906]: DEBUG nova.network.neutron [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 741.606669] env[61906]: INFO nova.compute.manager [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] [instance: 9e665b58-69e7-4ab9-b109-ab27725e66cc] Took 1.02 seconds to deallocate network for instance. [ 741.704186] env[61906]: DEBUG nova.compute.manager [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 741.883273] env[61906]: DEBUG nova.network.neutron [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 742.047554] env[61906]: DEBUG nova.network.neutron [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.227073] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.553020] env[61906]: DEBUG oslo_concurrency.lockutils [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Releasing lock "refresh_cache-8d2459c0-279c-4ea4-8cbc-6c69cfa35b27" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.553307] env[61906]: DEBUG nova.compute.manager [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 742.553484] env[61906]: DEBUG nova.compute.manager [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 742.553523] env[61906]: DEBUG nova.network.neutron [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 742.571957] env[61906]: DEBUG nova.network.neutron [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 742.634028] env[61906]: INFO nova.scheduler.client.report [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Deleted allocations for instance 9e665b58-69e7-4ab9-b109-ab27725e66cc [ 742.747052] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2316083d-0890-4a5c-9c46-41cb7b6c6c02 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.755955] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e1ac07c-7a7e-47f5-9a81-3a268e89c848 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.787746] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-936f5b1c-aabf-457d-800a-a4befc2d6af8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.796882] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f37ea09-6720-4d32-93b8-9688a47bfeb3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.812066] env[61906]: DEBUG nova.compute.provider_tree [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.076804] env[61906]: DEBUG nova.network.neutron [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.148921] env[61906]: DEBUG oslo_concurrency.lockutils [None req-61ff53f7-b010-4ead-bc17-5361b5906a67 tempest-InstanceActionsV221TestJSON-1606556463 tempest-InstanceActionsV221TestJSON-1606556463-project-member] Lock "9e665b58-69e7-4ab9-b109-ab27725e66cc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.895s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.315552] env[61906]: DEBUG nova.scheduler.client.report [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 743.583137] env[61906]: INFO nova.compute.manager [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] [instance: 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27] Took 1.03 seconds to deallocate network for instance. [ 743.653898] env[61906]: DEBUG nova.compute.manager [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 743.822076] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.472s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.822546] env[61906]: DEBUG nova.compute.manager [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 743.825370] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.580s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.181399] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.337160] env[61906]: DEBUG nova.compute.utils [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 744.342666] env[61906]: DEBUG nova.compute.manager [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 744.342666] env[61906]: DEBUG nova.network.neutron [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 744.438451] env[61906]: DEBUG nova.policy [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0de9abfd085f4a43a878da8448615bc4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e36ff7a500ef444cbf5a168c5a48208b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 744.625109] env[61906]: INFO nova.scheduler.client.report [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Deleted allocations for instance 8d2459c0-279c-4ea4-8cbc-6c69cfa35b27 [ 744.767286] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afdcdeb5-8db6-4d84-a9a2-061863e0cf64 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.773738] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf56a9c-91cd-4fb8-a82a-7378759fafd2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.812160] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22f213b7-5ab5-4316-9f42-370efa6fefec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.819443] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f64daad3-fcf5-4248-b842-29a1d233b4d5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.832620] env[61906]: DEBUG nova.compute.provider_tree [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 744.845036] env[61906]: DEBUG nova.compute.manager [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 745.137070] env[61906]: DEBUG oslo_concurrency.lockutils [None req-965577c2-17ff-4372-9feb-27a272dd0073 tempest-FloatingIPsAssociationNegativeTestJSON-685890679 tempest-FloatingIPsAssociationNegativeTestJSON-685890679-project-member] Lock "8d2459c0-279c-4ea4-8cbc-6c69cfa35b27" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.559s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.155888] env[61906]: DEBUG nova.network.neutron [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Successfully created port: c4e2007f-64ed-4304-bece-c5aa5d7c7b53 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 745.335578] env[61906]: DEBUG nova.scheduler.client.report [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 745.640132] env[61906]: DEBUG nova.compute.manager [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 745.841941] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.016s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.842626] env[61906]: ERROR nova.compute.manager [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2281c028-9a73-46e9-be06-40aeccecd788, please check neutron logs for more information. [ 745.842626] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Traceback (most recent call last): [ 745.842626] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 745.842626] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] self.driver.spawn(context, instance, image_meta, [ 745.842626] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 745.842626] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] self._vmops.spawn(context, instance, image_meta, injected_files, [ 745.842626] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 745.842626] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] vm_ref = self.build_virtual_machine(instance, [ 745.842626] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 745.842626] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] vif_infos = vmwarevif.get_vif_info(self._session, [ 745.842626] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 745.843070] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] for vif in network_info: [ 745.843070] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 745.843070] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] return self._sync_wrapper(fn, *args, **kwargs) [ 745.843070] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 745.843070] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] self.wait() [ 745.843070] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 745.843070] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] self[:] = self._gt.wait() [ 745.843070] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 745.843070] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] return self._exit_event.wait() [ 745.843070] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 745.843070] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] result = hub.switch() [ 745.843070] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 745.843070] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] return self.greenlet.switch() [ 745.843493] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 745.843493] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] result = function(*args, **kwargs) [ 745.843493] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 745.843493] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] return func(*args, **kwargs) [ 745.843493] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 745.843493] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] raise e [ 745.843493] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 745.843493] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] nwinfo = self.network_api.allocate_for_instance( [ 745.843493] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 745.843493] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] created_port_ids = self._update_ports_for_instance( [ 745.843493] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 745.843493] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] with excutils.save_and_reraise_exception(): [ 745.843493] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 745.843905] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] self.force_reraise() [ 745.843905] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 745.843905] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] raise self.value [ 745.843905] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 745.843905] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] updated_port = self._update_port( [ 745.843905] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 745.843905] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] _ensure_no_port_binding_failure(port) [ 745.843905] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 745.843905] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] raise exception.PortBindingFailed(port_id=port['id']) [ 745.843905] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] nova.exception.PortBindingFailed: Binding failed for port 2281c028-9a73-46e9-be06-40aeccecd788, please check neutron logs for more information. [ 745.843905] env[61906]: ERROR nova.compute.manager [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] [ 745.844296] env[61906]: DEBUG nova.compute.utils [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Binding failed for port 2281c028-9a73-46e9-be06-40aeccecd788, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 745.846923] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.232s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.852115] env[61906]: DEBUG nova.compute.manager [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Build of instance 7b4a9ebc-47fb-4078-969a-07642284b916 was re-scheduled: Binding failed for port 2281c028-9a73-46e9-be06-40aeccecd788, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 745.852638] env[61906]: DEBUG nova.compute.manager [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 745.853877] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Acquiring lock "refresh_cache-7b4a9ebc-47fb-4078-969a-07642284b916" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.853877] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Acquired lock "refresh_cache-7b4a9ebc-47fb-4078-969a-07642284b916" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.853877] env[61906]: DEBUG nova.network.neutron [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 745.865786] env[61906]: DEBUG nova.compute.manager [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 745.907013] env[61906]: DEBUG nova.virt.hardware [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 745.910050] env[61906]: DEBUG nova.virt.hardware [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 745.910244] env[61906]: DEBUG nova.virt.hardware [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 745.910436] env[61906]: DEBUG nova.virt.hardware [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 745.910584] env[61906]: DEBUG nova.virt.hardware [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 745.911011] env[61906]: DEBUG nova.virt.hardware [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 745.911239] env[61906]: DEBUG nova.virt.hardware [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 745.911406] env[61906]: DEBUG nova.virt.hardware [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 745.911576] env[61906]: DEBUG nova.virt.hardware [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 745.911741] env[61906]: DEBUG nova.virt.hardware [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 745.911916] env[61906]: DEBUG nova.virt.hardware [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 745.914321] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c7688b2-282a-4411-a600-ff768c477891 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.926828] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5702c4f-3aa0-4bee-93ad-66f28ccbe110 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.197928] env[61906]: DEBUG oslo_concurrency.lockutils [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.406964] env[61906]: DEBUG nova.network.neutron [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.603921] env[61906]: DEBUG nova.network.neutron [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.769422] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d35806-13c4-4da9-b785-a2a4657a2bed {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.780268] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e602e067-83e4-419b-b381-99865a42e579 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.810997] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95901387-e7f1-4fbf-af3c-1080a5ebdbbf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.820315] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8fc5b90-d072-44c8-a8e8-30e07d92bece {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.838492] env[61906]: DEBUG nova.compute.provider_tree [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.939881] env[61906]: DEBUG nova.compute.manager [req-c134ad76-3cf3-4c0c-84c7-2cdf3480ecba req-bfdebdf3-c129-47b4-9fb1-7e88cfa4c591 service nova] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Received event network-changed-c4e2007f-64ed-4304-bece-c5aa5d7c7b53 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 746.939881] env[61906]: DEBUG nova.compute.manager [req-c134ad76-3cf3-4c0c-84c7-2cdf3480ecba req-bfdebdf3-c129-47b4-9fb1-7e88cfa4c591 service nova] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Refreshing instance network info cache due to event network-changed-c4e2007f-64ed-4304-bece-c5aa5d7c7b53. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 746.940150] env[61906]: DEBUG oslo_concurrency.lockutils [req-c134ad76-3cf3-4c0c-84c7-2cdf3480ecba req-bfdebdf3-c129-47b4-9fb1-7e88cfa4c591 service nova] Acquiring lock "refresh_cache-adcd5674-ce1e-4b73-85bd-4344b5633c94" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.940267] env[61906]: DEBUG oslo_concurrency.lockutils [req-c134ad76-3cf3-4c0c-84c7-2cdf3480ecba req-bfdebdf3-c129-47b4-9fb1-7e88cfa4c591 service nova] Acquired lock "refresh_cache-adcd5674-ce1e-4b73-85bd-4344b5633c94" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.940427] env[61906]: DEBUG nova.network.neutron [req-c134ad76-3cf3-4c0c-84c7-2cdf3480ecba req-bfdebdf3-c129-47b4-9fb1-7e88cfa4c591 service nova] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Refreshing network info cache for port c4e2007f-64ed-4304-bece-c5aa5d7c7b53 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 747.109180] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Releasing lock "refresh_cache-7b4a9ebc-47fb-4078-969a-07642284b916" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.109540] env[61906]: DEBUG nova.compute.manager [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 747.110033] env[61906]: DEBUG nova.compute.manager [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 747.110033] env[61906]: DEBUG nova.network.neutron [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 747.145801] env[61906]: DEBUG nova.network.neutron [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.202414] env[61906]: ERROR nova.compute.manager [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c4e2007f-64ed-4304-bece-c5aa5d7c7b53, please check neutron logs for more information. [ 747.202414] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 747.202414] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 747.202414] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 747.202414] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 747.202414] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 747.202414] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 747.202414] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 747.202414] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.202414] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 747.202414] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.202414] env[61906]: ERROR nova.compute.manager raise self.value [ 747.202414] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 747.202414] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 747.202414] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.202414] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 747.203205] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.203205] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 747.203205] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c4e2007f-64ed-4304-bece-c5aa5d7c7b53, please check neutron logs for more information. [ 747.203205] env[61906]: ERROR nova.compute.manager [ 747.203205] env[61906]: Traceback (most recent call last): [ 747.203205] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 747.203205] env[61906]: listener.cb(fileno) [ 747.203205] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 747.203205] env[61906]: result = function(*args, **kwargs) [ 747.203205] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 747.203205] env[61906]: return func(*args, **kwargs) [ 747.203205] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 747.203205] env[61906]: raise e [ 747.203205] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 747.203205] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 747.203205] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 747.203205] env[61906]: created_port_ids = self._update_ports_for_instance( [ 747.203205] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 747.203205] env[61906]: with excutils.save_and_reraise_exception(): [ 747.203205] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.203205] env[61906]: self.force_reraise() [ 747.203205] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.203205] env[61906]: raise self.value [ 747.203205] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 747.203205] env[61906]: updated_port = self._update_port( [ 747.203205] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.203205] env[61906]: _ensure_no_port_binding_failure(port) [ 747.203205] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.203205] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 747.204821] env[61906]: nova.exception.PortBindingFailed: Binding failed for port c4e2007f-64ed-4304-bece-c5aa5d7c7b53, please check neutron logs for more information. [ 747.204821] env[61906]: Removing descriptor: 16 [ 747.204821] env[61906]: ERROR nova.compute.manager [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c4e2007f-64ed-4304-bece-c5aa5d7c7b53, please check neutron logs for more information. [ 747.204821] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Traceback (most recent call last): [ 747.204821] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 747.204821] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] yield resources [ 747.204821] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 747.204821] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] self.driver.spawn(context, instance, image_meta, [ 747.204821] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 747.204821] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] self._vmops.spawn(context, instance, image_meta, injected_files, [ 747.204821] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 747.204821] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] vm_ref = self.build_virtual_machine(instance, [ 747.205244] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 747.205244] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] vif_infos = vmwarevif.get_vif_info(self._session, [ 747.205244] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 747.205244] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] for vif in network_info: [ 747.205244] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 747.205244] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] return self._sync_wrapper(fn, *args, **kwargs) [ 747.205244] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 747.205244] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] self.wait() [ 747.205244] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 747.205244] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] self[:] = self._gt.wait() [ 747.205244] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 747.205244] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] return self._exit_event.wait() [ 747.205244] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 747.205890] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] result = hub.switch() [ 747.205890] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 747.205890] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] return self.greenlet.switch() [ 747.205890] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 747.205890] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] result = function(*args, **kwargs) [ 747.205890] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 747.205890] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] return func(*args, **kwargs) [ 747.205890] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 747.205890] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] raise e [ 747.205890] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 747.205890] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] nwinfo = self.network_api.allocate_for_instance( [ 747.205890] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 747.205890] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] created_port_ids = self._update_ports_for_instance( [ 747.206588] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 747.206588] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] with excutils.save_and_reraise_exception(): [ 747.206588] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.206588] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] self.force_reraise() [ 747.206588] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.206588] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] raise self.value [ 747.206588] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 747.206588] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] updated_port = self._update_port( [ 747.206588] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.206588] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] _ensure_no_port_binding_failure(port) [ 747.206588] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.206588] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] raise exception.PortBindingFailed(port_id=port['id']) [ 747.207153] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] nova.exception.PortBindingFailed: Binding failed for port c4e2007f-64ed-4304-bece-c5aa5d7c7b53, please check neutron logs for more information. [ 747.207153] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] [ 747.207153] env[61906]: INFO nova.compute.manager [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Terminating instance [ 747.207153] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "refresh_cache-adcd5674-ce1e-4b73-85bd-4344b5633c94" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.341013] env[61906]: DEBUG nova.scheduler.client.report [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 747.460230] env[61906]: DEBUG nova.network.neutron [req-c134ad76-3cf3-4c0c-84c7-2cdf3480ecba req-bfdebdf3-c129-47b4-9fb1-7e88cfa4c591 service nova] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.569533] env[61906]: DEBUG nova.network.neutron [req-c134ad76-3cf3-4c0c-84c7-2cdf3480ecba req-bfdebdf3-c129-47b4-9fb1-7e88cfa4c591 service nova] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.649284] env[61906]: DEBUG nova.network.neutron [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.853044] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.003s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.853044] env[61906]: ERROR nova.compute.manager [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6646962d-b310-45c7-924b-fcd70fcfe261, please check neutron logs for more information. [ 747.853044] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Traceback (most recent call last): [ 747.853044] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 747.853044] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] self.driver.spawn(context, instance, image_meta, [ 747.853044] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 747.853044] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 747.853044] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 747.853044] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] vm_ref = self.build_virtual_machine(instance, [ 747.853573] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 747.853573] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] vif_infos = vmwarevif.get_vif_info(self._session, [ 747.853573] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 747.853573] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] for vif in network_info: [ 747.853573] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 747.853573] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] return self._sync_wrapper(fn, *args, **kwargs) [ 747.853573] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 747.853573] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] self.wait() [ 747.853573] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 747.853573] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] self[:] = self._gt.wait() [ 747.853573] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 747.853573] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] return self._exit_event.wait() [ 747.853573] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 747.854022] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] result = hub.switch() [ 747.854022] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 747.854022] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] return self.greenlet.switch() [ 747.854022] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 747.854022] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] result = function(*args, **kwargs) [ 747.854022] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 747.854022] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] return func(*args, **kwargs) [ 747.854022] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 747.854022] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] raise e [ 747.854022] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 747.854022] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] nwinfo = self.network_api.allocate_for_instance( [ 747.854022] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 747.854022] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] created_port_ids = self._update_ports_for_instance( [ 747.854462] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 747.854462] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] with excutils.save_and_reraise_exception(): [ 747.854462] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.854462] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] self.force_reraise() [ 747.854462] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.854462] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] raise self.value [ 747.854462] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 747.854462] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] updated_port = self._update_port( [ 747.854462] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.854462] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] _ensure_no_port_binding_failure(port) [ 747.854462] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.854462] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] raise exception.PortBindingFailed(port_id=port['id']) [ 747.854908] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] nova.exception.PortBindingFailed: Binding failed for port 6646962d-b310-45c7-924b-fcd70fcfe261, please check neutron logs for more information. [ 747.854908] env[61906]: ERROR nova.compute.manager [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] [ 747.854908] env[61906]: DEBUG nova.compute.utils [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Binding failed for port 6646962d-b310-45c7-924b-fcd70fcfe261, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 747.855645] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.854s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.857698] env[61906]: INFO nova.compute.claims [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 747.861275] env[61906]: DEBUG nova.compute.manager [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Build of instance 7c389914-47a6-4ed4-aad6-e8aeeefbed7c was re-scheduled: Binding failed for port 6646962d-b310-45c7-924b-fcd70fcfe261, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 747.862026] env[61906]: DEBUG nova.compute.manager [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 747.862026] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "refresh_cache-7c389914-47a6-4ed4-aad6-e8aeeefbed7c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.862026] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired lock "refresh_cache-7c389914-47a6-4ed4-aad6-e8aeeefbed7c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.862234] env[61906]: DEBUG nova.network.neutron [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 748.072024] env[61906]: DEBUG oslo_concurrency.lockutils [req-c134ad76-3cf3-4c0c-84c7-2cdf3480ecba req-bfdebdf3-c129-47b4-9fb1-7e88cfa4c591 service nova] Releasing lock "refresh_cache-adcd5674-ce1e-4b73-85bd-4344b5633c94" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.072775] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired lock "refresh_cache-adcd5674-ce1e-4b73-85bd-4344b5633c94" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.072993] env[61906]: DEBUG nova.network.neutron [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 748.158239] env[61906]: INFO nova.compute.manager [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7b4a9ebc-47fb-4078-969a-07642284b916] Took 1.05 seconds to deallocate network for instance. [ 748.163052] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Acquiring lock "a651bd36-b527-4743-bc6e-5534c629e409" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.163415] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Lock "a651bd36-b527-4743-bc6e-5534c629e409" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.390258] env[61906]: DEBUG nova.network.neutron [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.523035] env[61906]: DEBUG nova.network.neutron [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.612161] env[61906]: DEBUG nova.network.neutron [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.758172] env[61906]: DEBUG nova.network.neutron [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.965381] env[61906]: DEBUG nova.compute.manager [req-1a34b3a7-e4a9-405e-a400-c426c9c62671 req-1814cf26-f5b3-4d31-b2b2-b0f0078dc59a service nova] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Received event network-vif-deleted-c4e2007f-64ed-4304-bece-c5aa5d7c7b53 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 749.028013] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Releasing lock "refresh_cache-7c389914-47a6-4ed4-aad6-e8aeeefbed7c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.028013] env[61906]: DEBUG nova.compute.manager [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 749.028013] env[61906]: DEBUG nova.compute.manager [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 749.028347] env[61906]: DEBUG nova.network.neutron [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 749.061618] env[61906]: DEBUG nova.network.neutron [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.204674] env[61906]: INFO nova.scheduler.client.report [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Deleted allocations for instance 7b4a9ebc-47fb-4078-969a-07642284b916 [ 749.261288] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Releasing lock "refresh_cache-adcd5674-ce1e-4b73-85bd-4344b5633c94" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.261691] env[61906]: DEBUG nova.compute.manager [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 749.261872] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 749.262188] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-daf9b09c-1217-434f-ad4d-6fda7306acc0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.273054] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-047d48b1-8213-44bd-b5cf-0de699b6e2ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.301370] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance adcd5674-ce1e-4b73-85bd-4344b5633c94 could not be found. [ 749.301649] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 749.301838] env[61906]: INFO nova.compute.manager [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Took 0.04 seconds to destroy the instance on the hypervisor. [ 749.302113] env[61906]: DEBUG oslo.service.loopingcall [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 749.305510] env[61906]: DEBUG nova.compute.manager [-] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 749.305618] env[61906]: DEBUG nova.network.neutron [-] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 749.323162] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c92dc32c-92bf-4b98-a237-3befdd587422 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.330277] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe9005e-bfac-467f-a3d1-c74485dbe8ab {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.334812] env[61906]: DEBUG nova.network.neutron [-] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.368426] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-280b911f-f4ef-4ab2-bf12-905a9c848afb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.376165] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2cca556-fa72-4f51-9bf6-f0373225e115 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.391110] env[61906]: DEBUG nova.compute.provider_tree [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 749.564654] env[61906]: DEBUG nova.network.neutron [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.717039] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5523d2f7-c799-4f94-bb43-f66e59196d15 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Lock "7b4a9ebc-47fb-4078-969a-07642284b916" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 149.661s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.837220] env[61906]: DEBUG nova.network.neutron [-] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.898785] env[61906]: DEBUG nova.scheduler.client.report [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 750.068500] env[61906]: INFO nova.compute.manager [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 7c389914-47a6-4ed4-aad6-e8aeeefbed7c] Took 1.04 seconds to deallocate network for instance. [ 750.219436] env[61906]: DEBUG nova.compute.manager [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 750.345330] env[61906]: INFO nova.compute.manager [-] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Took 1.04 seconds to deallocate network for instance. [ 750.349340] env[61906]: DEBUG nova.compute.claims [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 750.349340] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.406969] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.551s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.408029] env[61906]: DEBUG nova.compute.manager [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 750.411362] env[61906]: DEBUG oslo_concurrency.lockutils [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.265s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.753106] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.918219] env[61906]: DEBUG nova.compute.utils [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 750.922764] env[61906]: DEBUG nova.compute.manager [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 750.922943] env[61906]: DEBUG nova.network.neutron [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 751.007250] env[61906]: DEBUG nova.policy [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dcb235af9f1448dc9e6d2b79ec85df25', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4a262e68b08740d5a20d7df610319b18', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 751.115606] env[61906]: INFO nova.scheduler.client.report [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Deleted allocations for instance 7c389914-47a6-4ed4-aad6-e8aeeefbed7c [ 751.340031] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99e6cf41-ca5f-4f3f-bbd6-4b81c2914ada {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.349061] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-176e21d5-b4aa-4d91-91d9-88a9cc597b03 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.391229] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05a801f9-c010-48ff-b74d-a81373d3f7cc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.397725] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Acquiring lock "976021e6-0e38-4295-bdd4-83f00a7b7be0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.397958] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Lock "976021e6-0e38-4295-bdd4-83f00a7b7be0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.403911] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-888bf0b1-6358-4f4a-adeb-6dd3a137cf22 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.418739] env[61906]: DEBUG nova.compute.provider_tree [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.427427] env[61906]: DEBUG nova.compute.manager [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 751.598099] env[61906]: DEBUG nova.network.neutron [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Successfully created port: 5859dd38-7ba2-41cd-8dc1-e52927958ad4 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 751.630321] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9896ef27-6796-46ac-9896-c5163c8f0a91 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "7c389914-47a6-4ed4-aad6-e8aeeefbed7c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.296s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.921973] env[61906]: DEBUG nova.scheduler.client.report [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 752.132618] env[61906]: DEBUG nova.compute.manager [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 752.435008] env[61906]: DEBUG oslo_concurrency.lockutils [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.022s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.436247] env[61906]: ERROR nova.compute.manager [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8aa8e05d-7e39-402e-b0a1-d5ca5fc4e630, please check neutron logs for more information. [ 752.436247] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Traceback (most recent call last): [ 752.436247] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 752.436247] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] self.driver.spawn(context, instance, image_meta, [ 752.436247] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 752.436247] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] self._vmops.spawn(context, instance, image_meta, injected_files, [ 752.436247] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 752.436247] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] vm_ref = self.build_virtual_machine(instance, [ 752.436247] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 752.436247] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] vif_infos = vmwarevif.get_vif_info(self._session, [ 752.436247] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 752.436669] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] for vif in network_info: [ 752.436669] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 752.436669] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] return self._sync_wrapper(fn, *args, **kwargs) [ 752.436669] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 752.436669] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] self.wait() [ 752.436669] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 752.436669] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] self[:] = self._gt.wait() [ 752.436669] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 752.436669] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] return self._exit_event.wait() [ 752.436669] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 752.436669] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] result = hub.switch() [ 752.436669] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 752.436669] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] return self.greenlet.switch() [ 752.437097] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 752.437097] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] result = function(*args, **kwargs) [ 752.437097] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 752.437097] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] return func(*args, **kwargs) [ 752.437097] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 752.437097] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] raise e [ 752.437097] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 752.437097] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] nwinfo = self.network_api.allocate_for_instance( [ 752.437097] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 752.437097] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] created_port_ids = self._update_ports_for_instance( [ 752.437097] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 752.437097] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] with excutils.save_and_reraise_exception(): [ 752.437097] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 752.437516] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] self.force_reraise() [ 752.437516] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 752.437516] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] raise self.value [ 752.437516] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 752.437516] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] updated_port = self._update_port( [ 752.437516] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 752.437516] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] _ensure_no_port_binding_failure(port) [ 752.437516] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 752.437516] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] raise exception.PortBindingFailed(port_id=port['id']) [ 752.437516] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] nova.exception.PortBindingFailed: Binding failed for port 8aa8e05d-7e39-402e-b0a1-d5ca5fc4e630, please check neutron logs for more information. [ 752.437516] env[61906]: ERROR nova.compute.manager [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] [ 752.438407] env[61906]: DEBUG nova.compute.utils [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Binding failed for port 8aa8e05d-7e39-402e-b0a1-d5ca5fc4e630, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 752.439735] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.105s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.445641] env[61906]: DEBUG nova.compute.manager [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 752.448569] env[61906]: DEBUG nova.compute.manager [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Build of instance c15bc75b-87dc-4602-81ea-dbebce1b7763 was re-scheduled: Binding failed for port 8aa8e05d-7e39-402e-b0a1-d5ca5fc4e630, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 752.449115] env[61906]: DEBUG nova.compute.manager [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 752.449390] env[61906]: DEBUG oslo_concurrency.lockutils [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Acquiring lock "refresh_cache-c15bc75b-87dc-4602-81ea-dbebce1b7763" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.449799] env[61906]: DEBUG oslo_concurrency.lockutils [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Acquired lock "refresh_cache-c15bc75b-87dc-4602-81ea-dbebce1b7763" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.450016] env[61906]: DEBUG nova.network.neutron [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 752.476557] env[61906]: DEBUG nova.virt.hardware [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 752.476874] env[61906]: DEBUG nova.virt.hardware [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 752.477367] env[61906]: DEBUG nova.virt.hardware [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 752.477640] env[61906]: DEBUG nova.virt.hardware [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 752.477869] env[61906]: DEBUG nova.virt.hardware [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 752.478068] env[61906]: DEBUG nova.virt.hardware [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 752.478326] env[61906]: DEBUG nova.virt.hardware [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 752.478528] env[61906]: DEBUG nova.virt.hardware [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 752.478741] env[61906]: DEBUG nova.virt.hardware [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 752.478942] env[61906]: DEBUG nova.virt.hardware [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 752.479200] env[61906]: DEBUG nova.virt.hardware [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 752.480557] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50ab6fb1-9468-4141-81c7-52815f08379e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.494474] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87872dc1-5a22-43e1-aa8e-0f8675ba8b31 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.667304] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.982383] env[61906]: DEBUG nova.network.neutron [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.015052] env[61906]: ERROR nova.compute.manager [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5859dd38-7ba2-41cd-8dc1-e52927958ad4, please check neutron logs for more information. [ 753.015052] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 753.015052] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.015052] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 753.015052] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.015052] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 753.015052] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.015052] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 753.015052] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.015052] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 753.015052] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.015052] env[61906]: ERROR nova.compute.manager raise self.value [ 753.015052] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.015052] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 753.015052] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.015052] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 753.015869] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.015869] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 753.015869] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5859dd38-7ba2-41cd-8dc1-e52927958ad4, please check neutron logs for more information. [ 753.015869] env[61906]: ERROR nova.compute.manager [ 753.015869] env[61906]: Traceback (most recent call last): [ 753.015869] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 753.015869] env[61906]: listener.cb(fileno) [ 753.015869] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.015869] env[61906]: result = function(*args, **kwargs) [ 753.015869] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 753.015869] env[61906]: return func(*args, **kwargs) [ 753.015869] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 753.015869] env[61906]: raise e [ 753.015869] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.015869] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 753.015869] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.015869] env[61906]: created_port_ids = self._update_ports_for_instance( [ 753.015869] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.015869] env[61906]: with excutils.save_and_reraise_exception(): [ 753.015869] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.015869] env[61906]: self.force_reraise() [ 753.015869] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.015869] env[61906]: raise self.value [ 753.015869] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.015869] env[61906]: updated_port = self._update_port( [ 753.015869] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.015869] env[61906]: _ensure_no_port_binding_failure(port) [ 753.015869] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.015869] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 753.016752] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 5859dd38-7ba2-41cd-8dc1-e52927958ad4, please check neutron logs for more information. [ 753.016752] env[61906]: Removing descriptor: 16 [ 753.016752] env[61906]: ERROR nova.compute.manager [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5859dd38-7ba2-41cd-8dc1-e52927958ad4, please check neutron logs for more information. [ 753.016752] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Traceback (most recent call last): [ 753.016752] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 753.016752] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] yield resources [ 753.016752] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 753.016752] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] self.driver.spawn(context, instance, image_meta, [ 753.016752] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 753.016752] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 753.016752] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 753.016752] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] vm_ref = self.build_virtual_machine(instance, [ 753.017132] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 753.017132] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] vif_infos = vmwarevif.get_vif_info(self._session, [ 753.017132] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 753.017132] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] for vif in network_info: [ 753.017132] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 753.017132] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] return self._sync_wrapper(fn, *args, **kwargs) [ 753.017132] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 753.017132] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] self.wait() [ 753.017132] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 753.017132] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] self[:] = self._gt.wait() [ 753.017132] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 753.017132] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] return self._exit_event.wait() [ 753.017132] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 753.017527] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] result = hub.switch() [ 753.017527] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 753.017527] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] return self.greenlet.switch() [ 753.017527] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.017527] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] result = function(*args, **kwargs) [ 753.017527] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 753.017527] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] return func(*args, **kwargs) [ 753.017527] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 753.017527] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] raise e [ 753.017527] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.017527] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] nwinfo = self.network_api.allocate_for_instance( [ 753.017527] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.017527] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] created_port_ids = self._update_ports_for_instance( [ 753.017978] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.017978] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] with excutils.save_and_reraise_exception(): [ 753.017978] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.017978] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] self.force_reraise() [ 753.017978] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.017978] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] raise self.value [ 753.017978] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.017978] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] updated_port = self._update_port( [ 753.017978] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.017978] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] _ensure_no_port_binding_failure(port) [ 753.017978] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.017978] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] raise exception.PortBindingFailed(port_id=port['id']) [ 753.018453] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] nova.exception.PortBindingFailed: Binding failed for port 5859dd38-7ba2-41cd-8dc1-e52927958ad4, please check neutron logs for more information. [ 753.018453] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] [ 753.018453] env[61906]: INFO nova.compute.manager [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Terminating instance [ 753.018453] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Acquiring lock "refresh_cache-bf056a94-9e9b-4296-b621-7e8a7338330b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.018586] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Acquired lock "refresh_cache-bf056a94-9e9b-4296-b621-7e8a7338330b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.019490] env[61906]: DEBUG nova.network.neutron [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 753.090960] env[61906]: DEBUG nova.network.neutron [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.272782] env[61906]: DEBUG nova.compute.manager [req-c2703aac-6f60-4873-a0d9-c1fa31a01437 req-bd0448f9-0851-4476-a9a9-fa04dca95cbf service nova] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Received event network-changed-5859dd38-7ba2-41cd-8dc1-e52927958ad4 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 753.276265] env[61906]: DEBUG nova.compute.manager [req-c2703aac-6f60-4873-a0d9-c1fa31a01437 req-bd0448f9-0851-4476-a9a9-fa04dca95cbf service nova] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Refreshing instance network info cache due to event network-changed-5859dd38-7ba2-41cd-8dc1-e52927958ad4. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 753.276265] env[61906]: DEBUG oslo_concurrency.lockutils [req-c2703aac-6f60-4873-a0d9-c1fa31a01437 req-bd0448f9-0851-4476-a9a9-fa04dca95cbf service nova] Acquiring lock "refresh_cache-bf056a94-9e9b-4296-b621-7e8a7338330b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.279357] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f6054b-ec70-466e-99fe-7b95122d50c4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.287770] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aee35881-21e6-4bf8-ba20-00337ce3c27b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.319588] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2867bcd7-e8e3-4915-9334-7c1e9090b794 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.326914] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b75f873d-6d33-42b2-90bd-d5ddf89be886 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.339974] env[61906]: DEBUG nova.compute.provider_tree [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 753.463781] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Acquiring lock "7e05fee8-842f-4e03-853a-58bac5a9ac0e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.464089] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Lock "7e05fee8-842f-4e03-853a-58bac5a9ac0e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.540452] env[61906]: DEBUG nova.network.neutron [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.596554] env[61906]: DEBUG oslo_concurrency.lockutils [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Releasing lock "refresh_cache-c15bc75b-87dc-4602-81ea-dbebce1b7763" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.596935] env[61906]: DEBUG nova.compute.manager [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 753.596935] env[61906]: DEBUG nova.compute.manager [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 753.597114] env[61906]: DEBUG nova.network.neutron [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 753.616292] env[61906]: DEBUG nova.network.neutron [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.642660] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "9f1253e6-4727-4dca-bad2-04f0c5424cdb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.642887] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "9f1253e6-4727-4dca-bad2-04f0c5424cdb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.647547] env[61906]: DEBUG nova.network.neutron [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.842935] env[61906]: DEBUG nova.scheduler.client.report [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 753.953805] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 753.953805] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 754.119598] env[61906]: DEBUG nova.network.neutron [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.150177] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Releasing lock "refresh_cache-bf056a94-9e9b-4296-b621-7e8a7338330b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.150639] env[61906]: DEBUG nova.compute.manager [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 754.150829] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 754.151152] env[61906]: DEBUG oslo_concurrency.lockutils [req-c2703aac-6f60-4873-a0d9-c1fa31a01437 req-bd0448f9-0851-4476-a9a9-fa04dca95cbf service nova] Acquired lock "refresh_cache-bf056a94-9e9b-4296-b621-7e8a7338330b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.151333] env[61906]: DEBUG nova.network.neutron [req-c2703aac-6f60-4873-a0d9-c1fa31a01437 req-bd0448f9-0851-4476-a9a9-fa04dca95cbf service nova] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Refreshing network info cache for port 5859dd38-7ba2-41cd-8dc1-e52927958ad4 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 754.152350] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-72a018d7-80ed-4899-8910-bd137c3e375e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.162393] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b46f515-39ed-4b19-8541-858a5e154986 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.183076] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bf056a94-9e9b-4296-b621-7e8a7338330b could not be found. [ 754.183324] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 754.183452] env[61906]: INFO nova.compute.manager [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 754.183692] env[61906]: DEBUG oslo.service.loopingcall [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 754.183890] env[61906]: DEBUG nova.compute.manager [-] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 754.184110] env[61906]: DEBUG nova.network.neutron [-] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 754.198186] env[61906]: DEBUG nova.network.neutron [-] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.347729] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.908s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.348343] env[61906]: ERROR nova.compute.manager [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 766bdfb7-99ba-4eea-a90f-0da913c62e82, please check neutron logs for more information. [ 754.348343] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Traceback (most recent call last): [ 754.348343] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 754.348343] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] self.driver.spawn(context, instance, image_meta, [ 754.348343] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 754.348343] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 754.348343] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 754.348343] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] vm_ref = self.build_virtual_machine(instance, [ 754.348343] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 754.348343] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] vif_infos = vmwarevif.get_vif_info(self._session, [ 754.348343] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 754.348747] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] for vif in network_info: [ 754.348747] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 754.348747] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] return self._sync_wrapper(fn, *args, **kwargs) [ 754.348747] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 754.348747] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] self.wait() [ 754.348747] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 754.348747] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] self[:] = self._gt.wait() [ 754.348747] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 754.348747] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] return self._exit_event.wait() [ 754.348747] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 754.348747] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] current.throw(*self._exc) [ 754.348747] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 754.348747] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] result = function(*args, **kwargs) [ 754.349206] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 754.349206] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] return func(*args, **kwargs) [ 754.349206] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 754.349206] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] raise e [ 754.349206] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 754.349206] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] nwinfo = self.network_api.allocate_for_instance( [ 754.349206] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 754.349206] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] created_port_ids = self._update_ports_for_instance( [ 754.349206] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 754.349206] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] with excutils.save_and_reraise_exception(): [ 754.349206] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 754.349206] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] self.force_reraise() [ 754.349206] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 754.349760] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] raise self.value [ 754.349760] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 754.349760] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] updated_port = self._update_port( [ 754.349760] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 754.349760] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] _ensure_no_port_binding_failure(port) [ 754.349760] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 754.349760] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] raise exception.PortBindingFailed(port_id=port['id']) [ 754.349760] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] nova.exception.PortBindingFailed: Binding failed for port 766bdfb7-99ba-4eea-a90f-0da913c62e82, please check neutron logs for more information. [ 754.349760] env[61906]: ERROR nova.compute.manager [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] [ 754.349760] env[61906]: DEBUG nova.compute.utils [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Binding failed for port 766bdfb7-99ba-4eea-a90f-0da913c62e82, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 754.351040] env[61906]: DEBUG oslo_concurrency.lockutils [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.503s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.354041] env[61906]: DEBUG nova.compute.manager [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Build of instance c9cb7335-8977-4ddf-9d0e-ff2166abf8e7 was re-scheduled: Binding failed for port 766bdfb7-99ba-4eea-a90f-0da913c62e82, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 754.354377] env[61906]: DEBUG nova.compute.manager [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 754.354603] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Acquiring lock "refresh_cache-c9cb7335-8977-4ddf-9d0e-ff2166abf8e7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.354772] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Acquired lock "refresh_cache-c9cb7335-8977-4ddf-9d0e-ff2166abf8e7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.354899] env[61906]: DEBUG nova.network.neutron [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 754.460137] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 754.461116] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Starting heal instance info cache {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 754.461116] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Rebuilding the list of instances to heal {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 754.622597] env[61906]: INFO nova.compute.manager [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] [instance: c15bc75b-87dc-4602-81ea-dbebce1b7763] Took 1.03 seconds to deallocate network for instance. [ 754.671749] env[61906]: DEBUG nova.network.neutron [req-c2703aac-6f60-4873-a0d9-c1fa31a01437 req-bd0448f9-0851-4476-a9a9-fa04dca95cbf service nova] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.700466] env[61906]: DEBUG nova.network.neutron [-] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.795654] env[61906]: DEBUG nova.network.neutron [req-c2703aac-6f60-4873-a0d9-c1fa31a01437 req-bd0448f9-0851-4476-a9a9-fa04dca95cbf service nova] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.881030] env[61906]: DEBUG nova.network.neutron [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.965679] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 754.965920] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 754.965991] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Didn't find any instances for network info cache update. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 754.966858] env[61906]: DEBUG nova.network.neutron [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.967942] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 754.968255] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 754.968430] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 754.968586] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 754.968724] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 754.968866] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 754.968998] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61906) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 754.969658] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 755.193669] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18302a9f-a3e0-412f-ae6f-909352f51d77 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.201637] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb37679f-c726-4d44-b16f-c1fd959933c4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.206272] env[61906]: INFO nova.compute.manager [-] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Took 1.02 seconds to deallocate network for instance. [ 755.232433] env[61906]: DEBUG nova.compute.claims [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 755.232618] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.233350] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-599573bf-e372-473d-ad9d-62b0909e88c9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.240388] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-170d16d8-19dd-4049-92a8-f582d92db4d7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.253468] env[61906]: DEBUG nova.compute.provider_tree [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 755.298793] env[61906]: DEBUG oslo_concurrency.lockutils [req-c2703aac-6f60-4873-a0d9-c1fa31a01437 req-bd0448f9-0851-4476-a9a9-fa04dca95cbf service nova] Releasing lock "refresh_cache-bf056a94-9e9b-4296-b621-7e8a7338330b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.304283] env[61906]: DEBUG nova.compute.manager [req-fd9ace42-e77d-49c3-9cb8-4e24196a94c5 req-e215187f-6984-46de-bc46-cdd3493a53f2 service nova] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Received event network-vif-deleted-5859dd38-7ba2-41cd-8dc1-e52927958ad4 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 755.469625] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Releasing lock "refresh_cache-c9cb7335-8977-4ddf-9d0e-ff2166abf8e7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.469927] env[61906]: DEBUG nova.compute.manager [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 755.470070] env[61906]: DEBUG nova.compute.manager [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 755.470243] env[61906]: DEBUG nova.network.neutron [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 755.474170] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.487901] env[61906]: DEBUG nova.network.neutron [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 755.652894] env[61906]: INFO nova.scheduler.client.report [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Deleted allocations for instance c15bc75b-87dc-4602-81ea-dbebce1b7763 [ 755.759056] env[61906]: DEBUG nova.scheduler.client.report [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 755.990539] env[61906]: DEBUG nova.network.neutron [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.161152] env[61906]: DEBUG oslo_concurrency.lockutils [None req-411077e6-e727-4295-84ec-0f5a0e9a7048 tempest-VolumesAdminNegativeTest-1083297353 tempest-VolumesAdminNegativeTest-1083297353-project-member] Lock "c15bc75b-87dc-4602-81ea-dbebce1b7763" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.639s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.267022] env[61906]: DEBUG oslo_concurrency.lockutils [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.915s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.267022] env[61906]: ERROR nova.compute.manager [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f92e1b5b-fa31-476c-84aa-eade9006bbd2, please check neutron logs for more information. [ 756.267022] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Traceback (most recent call last): [ 756.267022] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 756.267022] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] self.driver.spawn(context, instance, image_meta, [ 756.267022] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 756.267022] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] self._vmops.spawn(context, instance, image_meta, injected_files, [ 756.267022] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 756.267022] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] vm_ref = self.build_virtual_machine(instance, [ 756.270122] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 756.270122] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] vif_infos = vmwarevif.get_vif_info(self._session, [ 756.270122] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 756.270122] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] for vif in network_info: [ 756.270122] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 756.270122] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] return self._sync_wrapper(fn, *args, **kwargs) [ 756.270122] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 756.270122] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] self.wait() [ 756.270122] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 756.270122] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] self[:] = self._gt.wait() [ 756.270122] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 756.270122] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] return self._exit_event.wait() [ 756.270122] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 756.271027] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] current.throw(*self._exc) [ 756.271027] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 756.271027] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] result = function(*args, **kwargs) [ 756.271027] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 756.271027] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] return func(*args, **kwargs) [ 756.271027] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 756.271027] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] raise e [ 756.271027] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 756.271027] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] nwinfo = self.network_api.allocate_for_instance( [ 756.271027] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 756.271027] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] created_port_ids = self._update_ports_for_instance( [ 756.271027] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 756.271027] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] with excutils.save_and_reraise_exception(): [ 756.271619] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.271619] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] self.force_reraise() [ 756.271619] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.271619] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] raise self.value [ 756.271619] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 756.271619] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] updated_port = self._update_port( [ 756.271619] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.271619] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] _ensure_no_port_binding_failure(port) [ 756.271619] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.271619] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] raise exception.PortBindingFailed(port_id=port['id']) [ 756.271619] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] nova.exception.PortBindingFailed: Binding failed for port f92e1b5b-fa31-476c-84aa-eade9006bbd2, please check neutron logs for more information. [ 756.271619] env[61906]: ERROR nova.compute.manager [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] [ 756.273341] env[61906]: DEBUG nova.compute.utils [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Binding failed for port f92e1b5b-fa31-476c-84aa-eade9006bbd2, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 756.273341] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.043s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.273341] env[61906]: INFO nova.compute.claims [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 756.274780] env[61906]: DEBUG nova.compute.manager [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Build of instance c4b41c42-e01f-464a-9423-1c0fcca75586 was re-scheduled: Binding failed for port f92e1b5b-fa31-476c-84aa-eade9006bbd2, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 756.275211] env[61906]: DEBUG nova.compute.manager [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 756.275440] env[61906]: DEBUG oslo_concurrency.lockutils [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Acquiring lock "refresh_cache-c4b41c42-e01f-464a-9423-1c0fcca75586" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.275610] env[61906]: DEBUG oslo_concurrency.lockutils [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Acquired lock "refresh_cache-c4b41c42-e01f-464a-9423-1c0fcca75586" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.275739] env[61906]: DEBUG nova.network.neutron [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 756.493726] env[61906]: INFO nova.compute.manager [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: c9cb7335-8977-4ddf-9d0e-ff2166abf8e7] Took 1.02 seconds to deallocate network for instance. [ 756.665032] env[61906]: DEBUG nova.compute.manager [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 756.804316] env[61906]: DEBUG nova.network.neutron [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 756.985091] env[61906]: DEBUG nova.network.neutron [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.197253] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.487794] env[61906]: DEBUG oslo_concurrency.lockutils [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Releasing lock "refresh_cache-c4b41c42-e01f-464a-9423-1c0fcca75586" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.488128] env[61906]: DEBUG nova.compute.manager [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 757.488373] env[61906]: DEBUG nova.compute.manager [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 757.488559] env[61906]: DEBUG nova.network.neutron [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 757.507056] env[61906]: DEBUG nova.network.neutron [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 757.522132] env[61906]: INFO nova.scheduler.client.report [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Deleted allocations for instance c9cb7335-8977-4ddf-9d0e-ff2166abf8e7 [ 757.642662] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a93042b8-ce14-493f-985e-3f9832b43894 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.655681] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfbe2b33-bcf3-4d27-ac70-d8bcdd57b492 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.686037] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1681641b-d6b8-430a-bee2-42de2d6797c6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.694518] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7dfcc91-a993-4812-9df1-37b0c63c1852 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.708087] env[61906]: DEBUG nova.compute.provider_tree [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.009620] env[61906]: DEBUG nova.network.neutron [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.037635] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a38d0de1-bfe8-4c9d-a2ad-6699ac2fbc39 tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Lock "c9cb7335-8977-4ddf-9d0e-ff2166abf8e7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 154.885s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.211925] env[61906]: DEBUG nova.scheduler.client.report [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 758.514925] env[61906]: INFO nova.compute.manager [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: c4b41c42-e01f-464a-9423-1c0fcca75586] Took 1.03 seconds to deallocate network for instance. [ 758.540922] env[61906]: DEBUG nova.compute.manager [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 758.717919] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.718488] env[61906]: DEBUG nova.compute.manager [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 758.723197] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.542s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.725464] env[61906]: INFO nova.compute.claims [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 759.070459] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.230505] env[61906]: DEBUG nova.compute.utils [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 759.234825] env[61906]: DEBUG nova.compute.manager [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 759.234825] env[61906]: DEBUG nova.network.neutron [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 759.291848] env[61906]: DEBUG nova.policy [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1b889ef55dbc4f4d9d88c0d57a8eaaaf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0983a7da8870442b86b96cfde74bf6a0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 759.551815] env[61906]: INFO nova.scheduler.client.report [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Deleted allocations for instance c4b41c42-e01f-464a-9423-1c0fcca75586 [ 759.639787] env[61906]: DEBUG nova.network.neutron [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Successfully created port: a41cd4c0-d7ec-48e3-9c57-a2d24241b560 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 759.735791] env[61906]: DEBUG nova.compute.manager [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 760.065693] env[61906]: DEBUG oslo_concurrency.lockutils [None req-26350414-980a-4c1f-9bc3-8f6be6f2ab87 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Lock "c4b41c42-e01f-464a-9423-1c0fcca75586" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.525s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.090169] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-badc7d59-8982-4f67-aa3a-70e1268c86f0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.098555] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a02f038-9703-4e9d-b16f-74b74d2de1df {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.131331] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c680778-77c1-44a0-9b9b-c88dcc178db6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.140136] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f0be187-b5c9-46af-a088-ee74116930a5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.154011] env[61906]: DEBUG nova.compute.provider_tree [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.528985] env[61906]: DEBUG nova.compute.manager [req-1af12bec-c9db-4bc3-803a-2f62d7fcd08d req-869814db-c552-42fc-a408-b2b9ab859581 service nova] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Received event network-changed-a41cd4c0-d7ec-48e3-9c57-a2d24241b560 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 760.529177] env[61906]: DEBUG nova.compute.manager [req-1af12bec-c9db-4bc3-803a-2f62d7fcd08d req-869814db-c552-42fc-a408-b2b9ab859581 service nova] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Refreshing instance network info cache due to event network-changed-a41cd4c0-d7ec-48e3-9c57-a2d24241b560. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 760.529397] env[61906]: DEBUG oslo_concurrency.lockutils [req-1af12bec-c9db-4bc3-803a-2f62d7fcd08d req-869814db-c552-42fc-a408-b2b9ab859581 service nova] Acquiring lock "refresh_cache-b071a9b5-3416-4898-9ae2-ef62186c1843" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.529566] env[61906]: DEBUG oslo_concurrency.lockutils [req-1af12bec-c9db-4bc3-803a-2f62d7fcd08d req-869814db-c552-42fc-a408-b2b9ab859581 service nova] Acquired lock "refresh_cache-b071a9b5-3416-4898-9ae2-ef62186c1843" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.529691] env[61906]: DEBUG nova.network.neutron [req-1af12bec-c9db-4bc3-803a-2f62d7fcd08d req-869814db-c552-42fc-a408-b2b9ab859581 service nova] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Refreshing network info cache for port a41cd4c0-d7ec-48e3-9c57-a2d24241b560 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 760.571619] env[61906]: DEBUG nova.compute.manager [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 760.665017] env[61906]: DEBUG nova.scheduler.client.report [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 760.751676] env[61906]: DEBUG nova.compute.manager [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 760.753970] env[61906]: ERROR nova.compute.manager [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a41cd4c0-d7ec-48e3-9c57-a2d24241b560, please check neutron logs for more information. [ 760.753970] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 760.753970] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 760.753970] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 760.753970] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 760.753970] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 760.753970] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 760.753970] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 760.753970] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.753970] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 760.753970] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.753970] env[61906]: ERROR nova.compute.manager raise self.value [ 760.753970] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 760.753970] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 760.753970] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.753970] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 760.754623] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.754623] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 760.754623] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a41cd4c0-d7ec-48e3-9c57-a2d24241b560, please check neutron logs for more information. [ 760.754623] env[61906]: ERROR nova.compute.manager [ 760.754623] env[61906]: Traceback (most recent call last): [ 760.754623] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 760.754623] env[61906]: listener.cb(fileno) [ 760.754623] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 760.754623] env[61906]: result = function(*args, **kwargs) [ 760.754623] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 760.754623] env[61906]: return func(*args, **kwargs) [ 760.754623] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 760.754623] env[61906]: raise e [ 760.754623] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 760.754623] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 760.754623] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 760.754623] env[61906]: created_port_ids = self._update_ports_for_instance( [ 760.754623] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 760.754623] env[61906]: with excutils.save_and_reraise_exception(): [ 760.754623] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.754623] env[61906]: self.force_reraise() [ 760.754623] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.754623] env[61906]: raise self.value [ 760.754623] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 760.754623] env[61906]: updated_port = self._update_port( [ 760.754623] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.754623] env[61906]: _ensure_no_port_binding_failure(port) [ 760.754623] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.754623] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 760.755523] env[61906]: nova.exception.PortBindingFailed: Binding failed for port a41cd4c0-d7ec-48e3-9c57-a2d24241b560, please check neutron logs for more information. [ 760.755523] env[61906]: Removing descriptor: 15 [ 760.780259] env[61906]: DEBUG nova.virt.hardware [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 760.780535] env[61906]: DEBUG nova.virt.hardware [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 760.780692] env[61906]: DEBUG nova.virt.hardware [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 760.780954] env[61906]: DEBUG nova.virt.hardware [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 760.781196] env[61906]: DEBUG nova.virt.hardware [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 760.781397] env[61906]: DEBUG nova.virt.hardware [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 760.781638] env[61906]: DEBUG nova.virt.hardware [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 760.781800] env[61906]: DEBUG nova.virt.hardware [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 760.781963] env[61906]: DEBUG nova.virt.hardware [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 760.782186] env[61906]: DEBUG nova.virt.hardware [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 760.782384] env[61906]: DEBUG nova.virt.hardware [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 760.783233] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29f782e6-85f3-42d5-86cf-e59649df27ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.791643] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27f9bd08-6f4a-46f1-b79d-851453f51cb2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.807671] env[61906]: ERROR nova.compute.manager [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a41cd4c0-d7ec-48e3-9c57-a2d24241b560, please check neutron logs for more information. [ 760.807671] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Traceback (most recent call last): [ 760.807671] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 760.807671] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] yield resources [ 760.807671] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 760.807671] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] self.driver.spawn(context, instance, image_meta, [ 760.807671] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 760.807671] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] self._vmops.spawn(context, instance, image_meta, injected_files, [ 760.807671] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 760.807671] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] vm_ref = self.build_virtual_machine(instance, [ 760.807671] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 760.808120] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] vif_infos = vmwarevif.get_vif_info(self._session, [ 760.808120] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 760.808120] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] for vif in network_info: [ 760.808120] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 760.808120] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] return self._sync_wrapper(fn, *args, **kwargs) [ 760.808120] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 760.808120] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] self.wait() [ 760.808120] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 760.808120] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] self[:] = self._gt.wait() [ 760.808120] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 760.808120] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] return self._exit_event.wait() [ 760.808120] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 760.808120] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] current.throw(*self._exc) [ 760.808542] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 760.808542] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] result = function(*args, **kwargs) [ 760.808542] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 760.808542] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] return func(*args, **kwargs) [ 760.808542] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 760.808542] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] raise e [ 760.808542] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 760.808542] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] nwinfo = self.network_api.allocate_for_instance( [ 760.808542] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 760.808542] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] created_port_ids = self._update_ports_for_instance( [ 760.808542] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 760.808542] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] with excutils.save_and_reraise_exception(): [ 760.808542] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.809063] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] self.force_reraise() [ 760.809063] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.809063] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] raise self.value [ 760.809063] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 760.809063] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] updated_port = self._update_port( [ 760.809063] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.809063] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] _ensure_no_port_binding_failure(port) [ 760.809063] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.809063] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] raise exception.PortBindingFailed(port_id=port['id']) [ 760.809063] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] nova.exception.PortBindingFailed: Binding failed for port a41cd4c0-d7ec-48e3-9c57-a2d24241b560, please check neutron logs for more information. [ 760.809063] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] [ 760.809063] env[61906]: INFO nova.compute.manager [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Terminating instance [ 760.810897] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Acquiring lock "refresh_cache-b071a9b5-3416-4898-9ae2-ef62186c1843" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.059213] env[61906]: DEBUG nova.network.neutron [req-1af12bec-c9db-4bc3-803a-2f62d7fcd08d req-869814db-c552-42fc-a408-b2b9ab859581 service nova] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.094252] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.167258] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.444s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.169574] env[61906]: DEBUG nova.compute.manager [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 761.170946] env[61906]: DEBUG oslo_concurrency.lockutils [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.973s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.172764] env[61906]: INFO nova.compute.claims [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 761.176048] env[61906]: DEBUG nova.network.neutron [req-1af12bec-c9db-4bc3-803a-2f62d7fcd08d req-869814db-c552-42fc-a408-b2b9ab859581 service nova] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.682040] env[61906]: DEBUG nova.compute.utils [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 761.682040] env[61906]: DEBUG oslo_concurrency.lockutils [req-1af12bec-c9db-4bc3-803a-2f62d7fcd08d req-869814db-c552-42fc-a408-b2b9ab859581 service nova] Releasing lock "refresh_cache-b071a9b5-3416-4898-9ae2-ef62186c1843" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.684112] env[61906]: DEBUG nova.compute.manager [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 761.684480] env[61906]: DEBUG nova.network.neutron [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 761.686387] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Acquired lock "refresh_cache-b071a9b5-3416-4898-9ae2-ef62186c1843" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.686694] env[61906]: DEBUG nova.network.neutron [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 761.806127] env[61906]: DEBUG nova.policy [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1b889ef55dbc4f4d9d88c0d57a8eaaaf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0983a7da8870442b86b96cfde74bf6a0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 762.187501] env[61906]: DEBUG nova.compute.manager [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 762.219186] env[61906]: DEBUG nova.network.neutron [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.313953] env[61906]: DEBUG nova.network.neutron [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.408132] env[61906]: DEBUG nova.network.neutron [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Successfully created port: 331a64dd-86e1-492c-aa13-2a0dd0b47a77 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 762.529768] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-226d95a1-b5ca-4738-98b9-59321c09036d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.537640] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53d4e53b-4bbe-4bd9-b4f0-8da4b34c92b2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.568216] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7799201a-cca5-4606-89ef-03225a7789d2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.575780] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c7e93c-f1bb-4a45-88b4-988242f64054 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.589132] env[61906]: DEBUG nova.compute.provider_tree [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.663761] env[61906]: DEBUG nova.compute.manager [req-30b92fd7-bcee-4417-bb39-f17b9cc78436 req-8881ae42-c66b-43d6-8221-1426110fa4e6 service nova] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Received event network-vif-deleted-a41cd4c0-d7ec-48e3-9c57-a2d24241b560 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 762.819871] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Releasing lock "refresh_cache-b071a9b5-3416-4898-9ae2-ef62186c1843" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.820320] env[61906]: DEBUG nova.compute.manager [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 762.820515] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 762.820808] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f4c4c9dc-6b9e-4536-9f23-4cd06438f187 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.830117] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39aa114d-7267-4106-8e0b-6bd6d5a05ff3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.850187] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b071a9b5-3416-4898-9ae2-ef62186c1843 could not be found. [ 762.850406] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 762.850585] env[61906]: INFO nova.compute.manager [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Took 0.03 seconds to destroy the instance on the hypervisor. [ 762.850856] env[61906]: DEBUG oslo.service.loopingcall [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 762.851079] env[61906]: DEBUG nova.compute.manager [-] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 762.851176] env[61906]: DEBUG nova.network.neutron [-] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 762.866606] env[61906]: DEBUG nova.network.neutron [-] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.092252] env[61906]: DEBUG nova.scheduler.client.report [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 763.195808] env[61906]: DEBUG nova.compute.manager [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 763.215588] env[61906]: DEBUG nova.virt.hardware [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:23Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 763.216500] env[61906]: DEBUG nova.virt.hardware [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 763.216781] env[61906]: DEBUG nova.virt.hardware [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 763.218016] env[61906]: DEBUG nova.virt.hardware [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 763.218016] env[61906]: DEBUG nova.virt.hardware [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 763.218016] env[61906]: DEBUG nova.virt.hardware [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 763.218016] env[61906]: DEBUG nova.virt.hardware [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 763.218016] env[61906]: DEBUG nova.virt.hardware [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 763.218284] env[61906]: DEBUG nova.virt.hardware [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 763.218558] env[61906]: DEBUG nova.virt.hardware [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 763.222104] env[61906]: DEBUG nova.virt.hardware [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 763.222104] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a4a14e7-8c20-4367-b0b1-345e27eb10d0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.229695] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-377cefcb-eeb5-4b31-810e-6396a97d1f95 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.369477] env[61906]: DEBUG nova.network.neutron [-] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.597531] env[61906]: DEBUG oslo_concurrency.lockutils [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.426s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.598095] env[61906]: DEBUG nova.compute.manager [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 763.601290] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.252s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.877732] env[61906]: INFO nova.compute.manager [-] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Took 1.02 seconds to deallocate network for instance. [ 763.878562] env[61906]: DEBUG nova.compute.claims [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 763.878906] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.108022] env[61906]: DEBUG nova.compute.utils [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 764.108022] env[61906]: DEBUG nova.compute.manager [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 764.108022] env[61906]: DEBUG nova.network.neutron [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 764.170052] env[61906]: DEBUG nova.policy [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b75fa0f1bb4c49329278f6513b99577d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '24b6a6da80d84cc5bb2c99051b92b731', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 764.273955] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Acquiring lock "80a6a0b8-17c3-46e8-8861-6d1759a19111" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.274176] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Lock "80a6a0b8-17c3-46e8-8861-6d1759a19111" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.407667] env[61906]: ERROR nova.compute.manager [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 331a64dd-86e1-492c-aa13-2a0dd0b47a77, please check neutron logs for more information. [ 764.407667] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 764.407667] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 764.407667] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 764.407667] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 764.407667] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 764.407667] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 764.407667] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 764.407667] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.407667] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 764.407667] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.407667] env[61906]: ERROR nova.compute.manager raise self.value [ 764.407667] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 764.407667] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 764.407667] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.407667] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 764.408232] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.408232] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 764.408232] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 331a64dd-86e1-492c-aa13-2a0dd0b47a77, please check neutron logs for more information. [ 764.408232] env[61906]: ERROR nova.compute.manager [ 764.408232] env[61906]: Traceback (most recent call last): [ 764.408232] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 764.408232] env[61906]: listener.cb(fileno) [ 764.408232] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 764.408232] env[61906]: result = function(*args, **kwargs) [ 764.408232] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 764.408232] env[61906]: return func(*args, **kwargs) [ 764.408232] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 764.408232] env[61906]: raise e [ 764.408232] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 764.408232] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 764.408232] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 764.408232] env[61906]: created_port_ids = self._update_ports_for_instance( [ 764.408232] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 764.408232] env[61906]: with excutils.save_and_reraise_exception(): [ 764.408232] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.408232] env[61906]: self.force_reraise() [ 764.408232] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.408232] env[61906]: raise self.value [ 764.408232] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 764.408232] env[61906]: updated_port = self._update_port( [ 764.408232] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.408232] env[61906]: _ensure_no_port_binding_failure(port) [ 764.408232] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.408232] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 764.409184] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 331a64dd-86e1-492c-aa13-2a0dd0b47a77, please check neutron logs for more information. [ 764.409184] env[61906]: Removing descriptor: 15 [ 764.409184] env[61906]: ERROR nova.compute.manager [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 331a64dd-86e1-492c-aa13-2a0dd0b47a77, please check neutron logs for more information. [ 764.409184] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Traceback (most recent call last): [ 764.409184] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 764.409184] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] yield resources [ 764.409184] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 764.409184] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] self.driver.spawn(context, instance, image_meta, [ 764.409184] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 764.409184] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] self._vmops.spawn(context, instance, image_meta, injected_files, [ 764.409184] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 764.409184] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] vm_ref = self.build_virtual_machine(instance, [ 764.409562] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 764.409562] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] vif_infos = vmwarevif.get_vif_info(self._session, [ 764.409562] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 764.409562] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] for vif in network_info: [ 764.409562] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 764.409562] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] return self._sync_wrapper(fn, *args, **kwargs) [ 764.409562] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 764.409562] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] self.wait() [ 764.409562] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 764.409562] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] self[:] = self._gt.wait() [ 764.409562] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 764.409562] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] return self._exit_event.wait() [ 764.409562] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 764.410049] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] result = hub.switch() [ 764.410049] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 764.410049] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] return self.greenlet.switch() [ 764.410049] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 764.410049] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] result = function(*args, **kwargs) [ 764.410049] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 764.410049] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] return func(*args, **kwargs) [ 764.410049] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 764.410049] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] raise e [ 764.410049] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 764.410049] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] nwinfo = self.network_api.allocate_for_instance( [ 764.410049] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 764.410049] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] created_port_ids = self._update_ports_for_instance( [ 764.410465] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 764.410465] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] with excutils.save_and_reraise_exception(): [ 764.410465] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.410465] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] self.force_reraise() [ 764.410465] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.410465] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] raise self.value [ 764.410465] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 764.410465] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] updated_port = self._update_port( [ 764.410465] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.410465] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] _ensure_no_port_binding_failure(port) [ 764.410465] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.410465] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] raise exception.PortBindingFailed(port_id=port['id']) [ 764.410859] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] nova.exception.PortBindingFailed: Binding failed for port 331a64dd-86e1-492c-aa13-2a0dd0b47a77, please check neutron logs for more information. [ 764.410859] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] [ 764.410859] env[61906]: INFO nova.compute.manager [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Terminating instance [ 764.411911] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Acquiring lock "refresh_cache-a07e49a5-e20a-49df-9326-a2e819934d15" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.412358] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Acquired lock "refresh_cache-a07e49a5-e20a-49df-9326-a2e819934d15" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.412358] env[61906]: DEBUG nova.network.neutron [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 764.498754] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-483c2d85-3d2b-4c82-8d4b-c737c167bf13 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.508424] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a73b5908-9aa5-4105-8ef7-d07c9440eb5e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.542883] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f44b4c3c-5cfe-494e-9862-baaa12fb51c5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.545407] env[61906]: DEBUG nova.network.neutron [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Successfully created port: 56526b59-0198-4803-b772-890468f44b15 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 764.552266] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2a82c75-d9e9-453a-85fe-17ade9be9cc3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.567283] env[61906]: DEBUG nova.compute.provider_tree [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 764.610848] env[61906]: DEBUG nova.compute.manager [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 764.800701] env[61906]: DEBUG nova.compute.manager [req-a269a0f5-7f19-4150-b64e-76bdf45fb8bb req-d51f44aa-2584-4183-b0a0-99594b9df00b service nova] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Received event network-changed-331a64dd-86e1-492c-aa13-2a0dd0b47a77 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 764.800701] env[61906]: DEBUG nova.compute.manager [req-a269a0f5-7f19-4150-b64e-76bdf45fb8bb req-d51f44aa-2584-4183-b0a0-99594b9df00b service nova] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Refreshing instance network info cache due to event network-changed-331a64dd-86e1-492c-aa13-2a0dd0b47a77. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 764.801166] env[61906]: DEBUG oslo_concurrency.lockutils [req-a269a0f5-7f19-4150-b64e-76bdf45fb8bb req-d51f44aa-2584-4183-b0a0-99594b9df00b service nova] Acquiring lock "refresh_cache-a07e49a5-e20a-49df-9326-a2e819934d15" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.933469] env[61906]: DEBUG nova.network.neutron [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.023898] env[61906]: DEBUG nova.network.neutron [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.069866] env[61906]: DEBUG nova.scheduler.client.report [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 765.526462] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Releasing lock "refresh_cache-a07e49a5-e20a-49df-9326-a2e819934d15" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.526903] env[61906]: DEBUG nova.compute.manager [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 765.527182] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 765.527410] env[61906]: DEBUG oslo_concurrency.lockutils [req-a269a0f5-7f19-4150-b64e-76bdf45fb8bb req-d51f44aa-2584-4183-b0a0-99594b9df00b service nova] Acquired lock "refresh_cache-a07e49a5-e20a-49df-9326-a2e819934d15" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.527576] env[61906]: DEBUG nova.network.neutron [req-a269a0f5-7f19-4150-b64e-76bdf45fb8bb req-d51f44aa-2584-4183-b0a0-99594b9df00b service nova] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Refreshing network info cache for port 331a64dd-86e1-492c-aa13-2a0dd0b47a77 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 765.528684] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7f396fc4-6451-4945-8451-1a6b37ca2ec2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.542939] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58832e1d-1272-4f33-900c-b5b95671d273 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.565939] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a07e49a5-e20a-49df-9326-a2e819934d15 could not be found. [ 765.566533] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 765.566533] env[61906]: INFO nova.compute.manager [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Took 0.04 seconds to destroy the instance on the hypervisor. [ 765.570020] env[61906]: DEBUG oslo.service.loopingcall [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 765.570020] env[61906]: DEBUG nova.compute.manager [-] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 765.570020] env[61906]: DEBUG nova.network.neutron [-] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 765.575169] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.974s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.575433] env[61906]: ERROR nova.compute.manager [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c4e2007f-64ed-4304-bece-c5aa5d7c7b53, please check neutron logs for more information. [ 765.575433] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Traceback (most recent call last): [ 765.575433] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 765.575433] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] self.driver.spawn(context, instance, image_meta, [ 765.575433] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 765.575433] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] self._vmops.spawn(context, instance, image_meta, injected_files, [ 765.575433] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 765.575433] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] vm_ref = self.build_virtual_machine(instance, [ 765.575433] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 765.575433] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] vif_infos = vmwarevif.get_vif_info(self._session, [ 765.575433] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 765.575906] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] for vif in network_info: [ 765.575906] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 765.575906] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] return self._sync_wrapper(fn, *args, **kwargs) [ 765.575906] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 765.575906] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] self.wait() [ 765.575906] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 765.575906] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] self[:] = self._gt.wait() [ 765.575906] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 765.575906] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] return self._exit_event.wait() [ 765.575906] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 765.575906] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] result = hub.switch() [ 765.575906] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 765.575906] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] return self.greenlet.switch() [ 765.576438] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 765.576438] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] result = function(*args, **kwargs) [ 765.576438] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 765.576438] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] return func(*args, **kwargs) [ 765.576438] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 765.576438] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] raise e [ 765.576438] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.576438] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] nwinfo = self.network_api.allocate_for_instance( [ 765.576438] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 765.576438] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] created_port_ids = self._update_ports_for_instance( [ 765.576438] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 765.576438] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] with excutils.save_and_reraise_exception(): [ 765.576438] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.576978] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] self.force_reraise() [ 765.576978] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.576978] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] raise self.value [ 765.576978] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 765.576978] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] updated_port = self._update_port( [ 765.576978] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.576978] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] _ensure_no_port_binding_failure(port) [ 765.576978] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.576978] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] raise exception.PortBindingFailed(port_id=port['id']) [ 765.576978] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] nova.exception.PortBindingFailed: Binding failed for port c4e2007f-64ed-4304-bece-c5aa5d7c7b53, please check neutron logs for more information. [ 765.576978] env[61906]: ERROR nova.compute.manager [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] [ 765.577426] env[61906]: DEBUG nova.compute.utils [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Binding failed for port c4e2007f-64ed-4304-bece-c5aa5d7c7b53, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 765.577426] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.824s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.578721] env[61906]: INFO nova.compute.claims [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 765.581572] env[61906]: DEBUG nova.compute.manager [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Build of instance adcd5674-ce1e-4b73-85bd-4344b5633c94 was re-scheduled: Binding failed for port c4e2007f-64ed-4304-bece-c5aa5d7c7b53, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 765.585523] env[61906]: DEBUG nova.compute.manager [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 765.585774] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "refresh_cache-adcd5674-ce1e-4b73-85bd-4344b5633c94" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.585924] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired lock "refresh_cache-adcd5674-ce1e-4b73-85bd-4344b5633c94" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.586109] env[61906]: DEBUG nova.network.neutron [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 765.598145] env[61906]: DEBUG nova.network.neutron [-] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.621675] env[61906]: DEBUG nova.compute.manager [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 765.657869] env[61906]: DEBUG nova.virt.hardware [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 765.658148] env[61906]: DEBUG nova.virt.hardware [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 765.658310] env[61906]: DEBUG nova.virt.hardware [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 765.658495] env[61906]: DEBUG nova.virt.hardware [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 765.658662] env[61906]: DEBUG nova.virt.hardware [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 765.658795] env[61906]: DEBUG nova.virt.hardware [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 765.658968] env[61906]: DEBUG nova.virt.hardware [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 765.659542] env[61906]: DEBUG nova.virt.hardware [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 765.659813] env[61906]: DEBUG nova.virt.hardware [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 765.660059] env[61906]: DEBUG nova.virt.hardware [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 765.660316] env[61906]: DEBUG nova.virt.hardware [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 765.661280] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9148192d-8268-4727-9a1f-d26a9ebb385f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.670858] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0690ae2a-611a-4d7c-a2bc-039d87293392 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.955955] env[61906]: DEBUG nova.compute.manager [req-7c5fa232-eb13-4e3d-af6b-10cafc09d191 req-33ed4773-4911-44ae-b2b8-238e64bb83cb service nova] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Received event network-changed-56526b59-0198-4803-b772-890468f44b15 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 765.955955] env[61906]: DEBUG nova.compute.manager [req-7c5fa232-eb13-4e3d-af6b-10cafc09d191 req-33ed4773-4911-44ae-b2b8-238e64bb83cb service nova] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Refreshing instance network info cache due to event network-changed-56526b59-0198-4803-b772-890468f44b15. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 765.955955] env[61906]: DEBUG oslo_concurrency.lockutils [req-7c5fa232-eb13-4e3d-af6b-10cafc09d191 req-33ed4773-4911-44ae-b2b8-238e64bb83cb service nova] Acquiring lock "refresh_cache-58b454d7-5370-499a-93dc-8613ee997b4d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.955955] env[61906]: DEBUG oslo_concurrency.lockutils [req-7c5fa232-eb13-4e3d-af6b-10cafc09d191 req-33ed4773-4911-44ae-b2b8-238e64bb83cb service nova] Acquired lock "refresh_cache-58b454d7-5370-499a-93dc-8613ee997b4d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.955955] env[61906]: DEBUG nova.network.neutron [req-7c5fa232-eb13-4e3d-af6b-10cafc09d191 req-33ed4773-4911-44ae-b2b8-238e64bb83cb service nova] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Refreshing network info cache for port 56526b59-0198-4803-b772-890468f44b15 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 766.065047] env[61906]: DEBUG nova.network.neutron [req-a269a0f5-7f19-4150-b64e-76bdf45fb8bb req-d51f44aa-2584-4183-b0a0-99594b9df00b service nova] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.100882] env[61906]: DEBUG nova.network.neutron [-] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.130279] env[61906]: DEBUG nova.network.neutron [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.158714] env[61906]: ERROR nova.compute.manager [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 56526b59-0198-4803-b772-890468f44b15, please check neutron logs for more information. [ 766.158714] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 766.158714] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 766.158714] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 766.158714] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 766.158714] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 766.158714] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 766.158714] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 766.158714] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.158714] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 766.158714] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.158714] env[61906]: ERROR nova.compute.manager raise self.value [ 766.158714] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 766.158714] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 766.158714] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.158714] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 766.159499] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.159499] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 766.159499] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 56526b59-0198-4803-b772-890468f44b15, please check neutron logs for more information. [ 766.159499] env[61906]: ERROR nova.compute.manager [ 766.159499] env[61906]: Traceback (most recent call last): [ 766.159499] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 766.159499] env[61906]: listener.cb(fileno) [ 766.159499] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 766.159499] env[61906]: result = function(*args, **kwargs) [ 766.159499] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 766.159499] env[61906]: return func(*args, **kwargs) [ 766.159499] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 766.159499] env[61906]: raise e [ 766.159499] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 766.159499] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 766.159499] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 766.159499] env[61906]: created_port_ids = self._update_ports_for_instance( [ 766.159499] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 766.159499] env[61906]: with excutils.save_and_reraise_exception(): [ 766.159499] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.159499] env[61906]: self.force_reraise() [ 766.159499] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.159499] env[61906]: raise self.value [ 766.159499] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 766.159499] env[61906]: updated_port = self._update_port( [ 766.159499] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.159499] env[61906]: _ensure_no_port_binding_failure(port) [ 766.159499] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.159499] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 766.160451] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 56526b59-0198-4803-b772-890468f44b15, please check neutron logs for more information. [ 766.160451] env[61906]: Removing descriptor: 16 [ 766.160451] env[61906]: ERROR nova.compute.manager [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 56526b59-0198-4803-b772-890468f44b15, please check neutron logs for more information. [ 766.160451] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Traceback (most recent call last): [ 766.160451] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 766.160451] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] yield resources [ 766.160451] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 766.160451] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] self.driver.spawn(context, instance, image_meta, [ 766.160451] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 766.160451] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 766.160451] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 766.160451] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] vm_ref = self.build_virtual_machine(instance, [ 766.160818] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 766.160818] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] vif_infos = vmwarevif.get_vif_info(self._session, [ 766.160818] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 766.160818] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] for vif in network_info: [ 766.160818] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 766.160818] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] return self._sync_wrapper(fn, *args, **kwargs) [ 766.160818] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 766.160818] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] self.wait() [ 766.160818] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 766.160818] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] self[:] = self._gt.wait() [ 766.160818] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 766.160818] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] return self._exit_event.wait() [ 766.160818] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 766.161264] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] result = hub.switch() [ 766.161264] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 766.161264] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] return self.greenlet.switch() [ 766.161264] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 766.161264] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] result = function(*args, **kwargs) [ 766.161264] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 766.161264] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] return func(*args, **kwargs) [ 766.161264] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 766.161264] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] raise e [ 766.161264] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 766.161264] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] nwinfo = self.network_api.allocate_for_instance( [ 766.161264] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 766.161264] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] created_port_ids = self._update_ports_for_instance( [ 766.161668] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 766.161668] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] with excutils.save_and_reraise_exception(): [ 766.161668] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.161668] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] self.force_reraise() [ 766.161668] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.161668] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] raise self.value [ 766.161668] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 766.161668] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] updated_port = self._update_port( [ 766.161668] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.161668] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] _ensure_no_port_binding_failure(port) [ 766.161668] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.161668] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] raise exception.PortBindingFailed(port_id=port['id']) [ 766.162531] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] nova.exception.PortBindingFailed: Binding failed for port 56526b59-0198-4803-b772-890468f44b15, please check neutron logs for more information. [ 766.162531] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] [ 766.162531] env[61906]: INFO nova.compute.manager [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Terminating instance [ 766.162531] env[61906]: DEBUG oslo_concurrency.lockutils [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Acquiring lock "refresh_cache-58b454d7-5370-499a-93dc-8613ee997b4d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.195844] env[61906]: DEBUG nova.network.neutron [req-a269a0f5-7f19-4150-b64e-76bdf45fb8bb req-d51f44aa-2584-4183-b0a0-99594b9df00b service nova] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.277784] env[61906]: DEBUG nova.network.neutron [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.477680] env[61906]: DEBUG nova.network.neutron [req-7c5fa232-eb13-4e3d-af6b-10cafc09d191 req-33ed4773-4911-44ae-b2b8-238e64bb83cb service nova] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.603555] env[61906]: DEBUG nova.network.neutron [req-7c5fa232-eb13-4e3d-af6b-10cafc09d191 req-33ed4773-4911-44ae-b2b8-238e64bb83cb service nova] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.604806] env[61906]: INFO nova.compute.manager [-] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Took 1.04 seconds to deallocate network for instance. [ 766.608792] env[61906]: DEBUG nova.compute.claims [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 766.608968] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.700483] env[61906]: DEBUG oslo_concurrency.lockutils [req-a269a0f5-7f19-4150-b64e-76bdf45fb8bb req-d51f44aa-2584-4183-b0a0-99594b9df00b service nova] Releasing lock "refresh_cache-a07e49a5-e20a-49df-9326-a2e819934d15" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.700736] env[61906]: DEBUG nova.compute.manager [req-a269a0f5-7f19-4150-b64e-76bdf45fb8bb req-d51f44aa-2584-4183-b0a0-99594b9df00b service nova] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Received event network-vif-deleted-331a64dd-86e1-492c-aa13-2a0dd0b47a77 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 766.778464] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Releasing lock "refresh_cache-adcd5674-ce1e-4b73-85bd-4344b5633c94" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.778692] env[61906]: DEBUG nova.compute.manager [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 766.778875] env[61906]: DEBUG nova.compute.manager [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 766.779070] env[61906]: DEBUG nova.network.neutron [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 766.800033] env[61906]: DEBUG nova.network.neutron [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.885846] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6cc82a7-08f8-43ee-8656-d68845248bee {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.893729] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d18e6ad2-f4d5-490f-b3c4-981a125bc519 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.925717] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5908809e-1028-4639-a84c-55662a7cc29a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.932767] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1973aa4c-e48a-4810-a059-15dc82413b74 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.945854] env[61906]: DEBUG nova.compute.provider_tree [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.106111] env[61906]: DEBUG oslo_concurrency.lockutils [req-7c5fa232-eb13-4e3d-af6b-10cafc09d191 req-33ed4773-4911-44ae-b2b8-238e64bb83cb service nova] Releasing lock "refresh_cache-58b454d7-5370-499a-93dc-8613ee997b4d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.106539] env[61906]: DEBUG oslo_concurrency.lockutils [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Acquired lock "refresh_cache-58b454d7-5370-499a-93dc-8613ee997b4d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.106721] env[61906]: DEBUG nova.network.neutron [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 767.304503] env[61906]: DEBUG nova.network.neutron [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.449534] env[61906]: DEBUG nova.scheduler.client.report [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 767.631747] env[61906]: DEBUG nova.network.neutron [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.774321] env[61906]: DEBUG nova.network.neutron [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.808502] env[61906]: INFO nova.compute.manager [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: adcd5674-ce1e-4b73-85bd-4344b5633c94] Took 1.03 seconds to deallocate network for instance. [ 767.955623] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.378s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.956153] env[61906]: DEBUG nova.compute.manager [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 767.958907] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.292s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.961155] env[61906]: INFO nova.compute.claims [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 767.991486] env[61906]: DEBUG nova.compute.manager [req-ac2e21ea-f87a-4c40-8774-19841c9d53fe req-b8d91068-4202-4e61-96d4-d91461d0690d service nova] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Received event network-vif-deleted-56526b59-0198-4803-b772-890468f44b15 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 768.276078] env[61906]: DEBUG oslo_concurrency.lockutils [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Releasing lock "refresh_cache-58b454d7-5370-499a-93dc-8613ee997b4d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.277476] env[61906]: DEBUG nova.compute.manager [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 768.277476] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 768.277476] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9092b98e-6dea-460c-a4c4-0310fa54262c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.287271] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-136cb232-7ae4-420f-8420-7194bd2d80d2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.309193] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 58b454d7-5370-499a-93dc-8613ee997b4d could not be found. [ 768.309471] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 768.309656] env[61906]: INFO nova.compute.manager [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 768.309908] env[61906]: DEBUG oslo.service.loopingcall [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 768.310285] env[61906]: DEBUG nova.compute.manager [-] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 768.310285] env[61906]: DEBUG nova.network.neutron [-] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 768.341282] env[61906]: DEBUG nova.network.neutron [-] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.468181] env[61906]: DEBUG nova.compute.utils [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 768.473488] env[61906]: DEBUG nova.compute.manager [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 768.473488] env[61906]: DEBUG nova.network.neutron [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 768.544389] env[61906]: DEBUG nova.policy [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b49e5a6aeb1d43e290a645e14861b889', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '53644a4143a24759a9ff2b5e28b84fb4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 768.842568] env[61906]: INFO nova.scheduler.client.report [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Deleted allocations for instance adcd5674-ce1e-4b73-85bd-4344b5633c94 [ 768.848018] env[61906]: DEBUG nova.network.neutron [-] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.932747] env[61906]: DEBUG nova.network.neutron [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Successfully created port: 40b6fa5b-eee9-46db-b7d0-e38d0315d3e5 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 768.971090] env[61906]: DEBUG nova.compute.manager [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 769.299293] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1153bf47-2155-4539-9c1e-7640c0099f2a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.307067] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ceb3dec-f514-421b-b6a9-175d1ddc4553 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.339016] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8be9bd85-6fd9-4e16-886c-30b40e051504 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.346298] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a01006-51f3-4c20-b200-eb170e679833 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.351543] env[61906]: INFO nova.compute.manager [-] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Took 1.04 seconds to deallocate network for instance. [ 769.351970] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2f371b6c-9599-4684-9c17-09648e3c783d tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "adcd5674-ce1e-4b73-85bd-4344b5633c94" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 165.639s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.362069] env[61906]: DEBUG nova.compute.provider_tree [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 769.366023] env[61906]: DEBUG nova.compute.claims [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 769.366023] env[61906]: DEBUG oslo_concurrency.lockutils [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.868228] env[61906]: DEBUG nova.compute.manager [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 769.871775] env[61906]: DEBUG nova.scheduler.client.report [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 769.885586] env[61906]: DEBUG nova.compute.manager [req-5fa7351f-a36f-4d2f-8a77-9b6bdec7e17d req-b62fdd5f-7643-4462-805b-af82a34ff4ba service nova] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Received event network-changed-40b6fa5b-eee9-46db-b7d0-e38d0315d3e5 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 769.885892] env[61906]: DEBUG nova.compute.manager [req-5fa7351f-a36f-4d2f-8a77-9b6bdec7e17d req-b62fdd5f-7643-4462-805b-af82a34ff4ba service nova] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Refreshing instance network info cache due to event network-changed-40b6fa5b-eee9-46db-b7d0-e38d0315d3e5. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 769.885892] env[61906]: DEBUG oslo_concurrency.lockutils [req-5fa7351f-a36f-4d2f-8a77-9b6bdec7e17d req-b62fdd5f-7643-4462-805b-af82a34ff4ba service nova] Acquiring lock "refresh_cache-3df44b70-ee72-40d7-9129-b5e94e121b0a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.886012] env[61906]: DEBUG oslo_concurrency.lockutils [req-5fa7351f-a36f-4d2f-8a77-9b6bdec7e17d req-b62fdd5f-7643-4462-805b-af82a34ff4ba service nova] Acquired lock "refresh_cache-3df44b70-ee72-40d7-9129-b5e94e121b0a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.886178] env[61906]: DEBUG nova.network.neutron [req-5fa7351f-a36f-4d2f-8a77-9b6bdec7e17d req-b62fdd5f-7643-4462-805b-af82a34ff4ba service nova] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Refreshing network info cache for port 40b6fa5b-eee9-46db-b7d0-e38d0315d3e5 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 769.933341] env[61906]: ERROR nova.compute.manager [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 40b6fa5b-eee9-46db-b7d0-e38d0315d3e5, please check neutron logs for more information. [ 769.933341] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 769.933341] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.933341] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 769.933341] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.933341] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 769.933341] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.933341] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 769.933341] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.933341] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 769.933341] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.933341] env[61906]: ERROR nova.compute.manager raise self.value [ 769.933341] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.933341] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 769.933341] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.933341] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 769.933983] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.933983] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 769.933983] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 40b6fa5b-eee9-46db-b7d0-e38d0315d3e5, please check neutron logs for more information. [ 769.933983] env[61906]: ERROR nova.compute.manager [ 769.933983] env[61906]: Traceback (most recent call last): [ 769.933983] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 769.933983] env[61906]: listener.cb(fileno) [ 769.933983] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.933983] env[61906]: result = function(*args, **kwargs) [ 769.933983] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 769.933983] env[61906]: return func(*args, **kwargs) [ 769.933983] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 769.933983] env[61906]: raise e [ 769.933983] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.933983] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 769.933983] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.933983] env[61906]: created_port_ids = self._update_ports_for_instance( [ 769.933983] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.933983] env[61906]: with excutils.save_and_reraise_exception(): [ 769.933983] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.933983] env[61906]: self.force_reraise() [ 769.933983] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.933983] env[61906]: raise self.value [ 769.933983] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.933983] env[61906]: updated_port = self._update_port( [ 769.933983] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.933983] env[61906]: _ensure_no_port_binding_failure(port) [ 769.933983] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.933983] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 769.934954] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 40b6fa5b-eee9-46db-b7d0-e38d0315d3e5, please check neutron logs for more information. [ 769.934954] env[61906]: Removing descriptor: 18 [ 769.984454] env[61906]: DEBUG nova.compute.manager [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 770.015687] env[61906]: DEBUG nova.virt.hardware [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 770.015926] env[61906]: DEBUG nova.virt.hardware [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 770.016112] env[61906]: DEBUG nova.virt.hardware [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 770.016306] env[61906]: DEBUG nova.virt.hardware [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 770.016458] env[61906]: DEBUG nova.virt.hardware [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 770.016605] env[61906]: DEBUG nova.virt.hardware [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 770.016816] env[61906]: DEBUG nova.virt.hardware [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 770.016974] env[61906]: DEBUG nova.virt.hardware [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 770.017158] env[61906]: DEBUG nova.virt.hardware [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 770.017323] env[61906]: DEBUG nova.virt.hardware [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 770.017498] env[61906]: DEBUG nova.virt.hardware [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 770.018427] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f341671-119a-4d2e-ad07-5f933828e2ab {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.027631] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df91449d-d465-416f-ba1e-532dc7a83808 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.046136] env[61906]: ERROR nova.compute.manager [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 40b6fa5b-eee9-46db-b7d0-e38d0315d3e5, please check neutron logs for more information. [ 770.046136] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Traceback (most recent call last): [ 770.046136] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 770.046136] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] yield resources [ 770.046136] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 770.046136] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] self.driver.spawn(context, instance, image_meta, [ 770.046136] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 770.046136] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 770.046136] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 770.046136] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] vm_ref = self.build_virtual_machine(instance, [ 770.046136] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 770.046974] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] vif_infos = vmwarevif.get_vif_info(self._session, [ 770.046974] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 770.046974] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] for vif in network_info: [ 770.046974] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 770.046974] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] return self._sync_wrapper(fn, *args, **kwargs) [ 770.046974] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 770.046974] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] self.wait() [ 770.046974] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 770.046974] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] self[:] = self._gt.wait() [ 770.046974] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 770.046974] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] return self._exit_event.wait() [ 770.046974] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 770.046974] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] current.throw(*self._exc) [ 770.047469] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 770.047469] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] result = function(*args, **kwargs) [ 770.047469] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 770.047469] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] return func(*args, **kwargs) [ 770.047469] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 770.047469] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] raise e [ 770.047469] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 770.047469] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] nwinfo = self.network_api.allocate_for_instance( [ 770.047469] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 770.047469] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] created_port_ids = self._update_ports_for_instance( [ 770.047469] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 770.047469] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] with excutils.save_and_reraise_exception(): [ 770.047469] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 770.048018] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] self.force_reraise() [ 770.048018] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 770.048018] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] raise self.value [ 770.048018] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 770.048018] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] updated_port = self._update_port( [ 770.048018] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 770.048018] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] _ensure_no_port_binding_failure(port) [ 770.048018] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 770.048018] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] raise exception.PortBindingFailed(port_id=port['id']) [ 770.048018] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] nova.exception.PortBindingFailed: Binding failed for port 40b6fa5b-eee9-46db-b7d0-e38d0315d3e5, please check neutron logs for more information. [ 770.048018] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] [ 770.048018] env[61906]: INFO nova.compute.manager [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Terminating instance [ 770.048800] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "refresh_cache-3df44b70-ee72-40d7-9129-b5e94e121b0a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.381332] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.421s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.381332] env[61906]: DEBUG nova.compute.manager [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 770.389144] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.155s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.409127] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.555818] env[61906]: DEBUG nova.network.neutron [req-5fa7351f-a36f-4d2f-8a77-9b6bdec7e17d req-b62fdd5f-7643-4462-805b-af82a34ff4ba service nova] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.732027] env[61906]: DEBUG nova.network.neutron [req-5fa7351f-a36f-4d2f-8a77-9b6bdec7e17d req-b62fdd5f-7643-4462-805b-af82a34ff4ba service nova] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.899575] env[61906]: DEBUG nova.compute.utils [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 770.901250] env[61906]: DEBUG nova.compute.manager [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 770.901455] env[61906]: DEBUG nova.network.neutron [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 771.023504] env[61906]: DEBUG nova.policy [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9f45d271db144a24b28cc07a1b2fb2d6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b04e99b97ceb4f59a034c0b6f36d5711', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 771.237062] env[61906]: DEBUG oslo_concurrency.lockutils [req-5fa7351f-a36f-4d2f-8a77-9b6bdec7e17d req-b62fdd5f-7643-4462-805b-af82a34ff4ba service nova] Releasing lock "refresh_cache-3df44b70-ee72-40d7-9129-b5e94e121b0a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.237815] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired lock "refresh_cache-3df44b70-ee72-40d7-9129-b5e94e121b0a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.238047] env[61906]: DEBUG nova.network.neutron [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 771.338787] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2de4447-3414-4301-bd11-7f56cb24c871 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.348254] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7125bfe-fa5b-4d49-8f6c-dbda201102f3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.381701] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33132ecc-24d7-4143-a9be-08bceadc42a9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.390619] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65e2c6e8-0a09-4b71-a03e-8e933eb211d3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.410660] env[61906]: DEBUG nova.compute.manager [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 771.413327] env[61906]: DEBUG nova.compute.provider_tree [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 771.454207] env[61906]: DEBUG nova.network.neutron [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Successfully created port: 955316c4-ca43-4931-a1f2-cf43d26ffa46 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 771.778595] env[61906]: DEBUG nova.network.neutron [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.923211] env[61906]: DEBUG nova.scheduler.client.report [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 771.947344] env[61906]: DEBUG nova.compute.manager [req-9249efe7-5000-4908-b9d8-82920b406f6a req-2f88f841-8525-46ad-90e2-9565bcada1d0 service nova] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Received event network-vif-deleted-40b6fa5b-eee9-46db-b7d0-e38d0315d3e5 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 771.980129] env[61906]: DEBUG nova.network.neutron [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.429988] env[61906]: DEBUG nova.compute.manager [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 772.433087] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.046s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.433702] env[61906]: ERROR nova.compute.manager [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5859dd38-7ba2-41cd-8dc1-e52927958ad4, please check neutron logs for more information. [ 772.433702] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Traceback (most recent call last): [ 772.433702] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 772.433702] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] self.driver.spawn(context, instance, image_meta, [ 772.433702] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 772.433702] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 772.433702] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 772.433702] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] vm_ref = self.build_virtual_machine(instance, [ 772.433702] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 772.433702] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] vif_infos = vmwarevif.get_vif_info(self._session, [ 772.433702] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 772.434297] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] for vif in network_info: [ 772.434297] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 772.434297] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] return self._sync_wrapper(fn, *args, **kwargs) [ 772.434297] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 772.434297] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] self.wait() [ 772.434297] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 772.434297] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] self[:] = self._gt.wait() [ 772.434297] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 772.434297] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] return self._exit_event.wait() [ 772.434297] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 772.434297] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] result = hub.switch() [ 772.434297] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 772.434297] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] return self.greenlet.switch() [ 772.434869] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 772.434869] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] result = function(*args, **kwargs) [ 772.434869] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 772.434869] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] return func(*args, **kwargs) [ 772.434869] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 772.434869] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] raise e [ 772.434869] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 772.434869] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] nwinfo = self.network_api.allocate_for_instance( [ 772.434869] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 772.434869] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] created_port_ids = self._update_ports_for_instance( [ 772.434869] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 772.434869] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] with excutils.save_and_reraise_exception(): [ 772.434869] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 772.435475] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] self.force_reraise() [ 772.435475] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 772.435475] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] raise self.value [ 772.435475] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 772.435475] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] updated_port = self._update_port( [ 772.435475] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 772.435475] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] _ensure_no_port_binding_failure(port) [ 772.435475] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 772.435475] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] raise exception.PortBindingFailed(port_id=port['id']) [ 772.435475] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] nova.exception.PortBindingFailed: Binding failed for port 5859dd38-7ba2-41cd-8dc1-e52927958ad4, please check neutron logs for more information. [ 772.435475] env[61906]: ERROR nova.compute.manager [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] [ 772.436028] env[61906]: DEBUG nova.compute.utils [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Binding failed for port 5859dd38-7ba2-41cd-8dc1-e52927958ad4, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 772.437707] env[61906]: DEBUG nova.compute.manager [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Build of instance bf056a94-9e9b-4296-b621-7e8a7338330b was re-scheduled: Binding failed for port 5859dd38-7ba2-41cd-8dc1-e52927958ad4, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 772.438093] env[61906]: DEBUG nova.compute.manager [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 772.438337] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Acquiring lock "refresh_cache-bf056a94-9e9b-4296-b621-7e8a7338330b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.438490] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Acquired lock "refresh_cache-bf056a94-9e9b-4296-b621-7e8a7338330b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.438652] env[61906]: DEBUG nova.network.neutron [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 772.439844] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.966s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.441249] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.441249] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61906) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 772.441249] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.243s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.442777] env[61906]: INFO nova.compute.claims [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 772.450583] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac7cd7fa-2c94-4134-a759-ce7f432b491c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.471795] env[61906]: DEBUG nova.virt.hardware [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 772.471795] env[61906]: DEBUG nova.virt.hardware [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 772.471795] env[61906]: DEBUG nova.virt.hardware [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 772.472045] env[61906]: DEBUG nova.virt.hardware [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 772.472045] env[61906]: DEBUG nova.virt.hardware [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 772.472045] env[61906]: DEBUG nova.virt.hardware [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 772.472045] env[61906]: DEBUG nova.virt.hardware [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 772.472045] env[61906]: DEBUG nova.virt.hardware [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 772.472344] env[61906]: DEBUG nova.virt.hardware [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 772.472344] env[61906]: DEBUG nova.virt.hardware [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 772.472344] env[61906]: DEBUG nova.virt.hardware [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 772.472799] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7486b7a1-aad0-45af-8d18-9c4927ad7868 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.477499] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9823c25-86b2-4dd0-a32a-91d429644be2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.485082] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Releasing lock "refresh_cache-3df44b70-ee72-40d7-9129-b5e94e121b0a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.485436] env[61906]: DEBUG nova.compute.manager [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 772.485751] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 772.493887] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d5b61f4f-7b6f-4b1c-ac20-29277880c099 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.496734] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ef0bc7-0462-4bd5-a438-3290126b61a7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.501576] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-353448f3-a0e9-4d28-a074-ef591450a2f9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.518925] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c69194f-0a7e-460c-9daf-e902c00c3c26 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.537956] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5527c757-00e6-4a57-a502-66bf1ba973b3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.545395] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3df44b70-ee72-40d7-9129-b5e94e121b0a could not be found. [ 772.545666] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 772.545940] env[61906]: INFO nova.compute.manager [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Took 0.06 seconds to destroy the instance on the hypervisor. [ 772.546240] env[61906]: DEBUG oslo.service.loopingcall [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 772.546829] env[61906]: DEBUG nova.compute.manager [-] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 772.546984] env[61906]: DEBUG nova.network.neutron [-] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 772.573021] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181435MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=61906) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 772.573255] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.574529] env[61906]: DEBUG nova.network.neutron [-] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 772.957369] env[61906]: ERROR nova.compute.manager [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 955316c4-ca43-4931-a1f2-cf43d26ffa46, please check neutron logs for more information. [ 772.957369] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 772.957369] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 772.957369] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 772.957369] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 772.957369] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 772.957369] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 772.957369] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 772.957369] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 772.957369] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 772.957369] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 772.957369] env[61906]: ERROR nova.compute.manager raise self.value [ 772.957369] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 772.957369] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 772.957369] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 772.957369] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 772.958531] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 772.958531] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 772.958531] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 955316c4-ca43-4931-a1f2-cf43d26ffa46, please check neutron logs for more information. [ 772.958531] env[61906]: ERROR nova.compute.manager [ 772.958531] env[61906]: Traceback (most recent call last): [ 772.958531] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 772.958531] env[61906]: listener.cb(fileno) [ 772.958531] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 772.958531] env[61906]: result = function(*args, **kwargs) [ 772.958531] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 772.958531] env[61906]: return func(*args, **kwargs) [ 772.958531] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 772.958531] env[61906]: raise e [ 772.958531] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 772.958531] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 772.958531] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 772.958531] env[61906]: created_port_ids = self._update_ports_for_instance( [ 772.958531] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 772.958531] env[61906]: with excutils.save_and_reraise_exception(): [ 772.958531] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 772.958531] env[61906]: self.force_reraise() [ 772.958531] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 772.958531] env[61906]: raise self.value [ 772.958531] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 772.958531] env[61906]: updated_port = self._update_port( [ 772.958531] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 772.958531] env[61906]: _ensure_no_port_binding_failure(port) [ 772.958531] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 772.958531] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 772.959915] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 955316c4-ca43-4931-a1f2-cf43d26ffa46, please check neutron logs for more information. [ 772.959915] env[61906]: Removing descriptor: 18 [ 772.959915] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Acquiring lock "d60575a6-af13-4d81-95d1-081d0f3e2ce6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.959915] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Lock "d60575a6-af13-4d81-95d1-081d0f3e2ce6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.959915] env[61906]: ERROR nova.compute.manager [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 955316c4-ca43-4931-a1f2-cf43d26ffa46, please check neutron logs for more information. [ 772.959915] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Traceback (most recent call last): [ 772.959915] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 772.959915] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] yield resources [ 772.960375] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 772.960375] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] self.driver.spawn(context, instance, image_meta, [ 772.960375] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 772.960375] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 772.960375] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 772.960375] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] vm_ref = self.build_virtual_machine(instance, [ 772.960375] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 772.960375] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] vif_infos = vmwarevif.get_vif_info(self._session, [ 772.960375] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 772.960375] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] for vif in network_info: [ 772.960375] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 772.960375] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] return self._sync_wrapper(fn, *args, **kwargs) [ 772.960375] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 772.960888] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] self.wait() [ 772.960888] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 772.960888] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] self[:] = self._gt.wait() [ 772.960888] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 772.960888] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] return self._exit_event.wait() [ 772.960888] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 772.960888] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] result = hub.switch() [ 772.960888] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 772.960888] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] return self.greenlet.switch() [ 772.960888] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 772.960888] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] result = function(*args, **kwargs) [ 772.960888] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 772.960888] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] return func(*args, **kwargs) [ 772.961359] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 772.961359] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] raise e [ 772.961359] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 772.961359] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] nwinfo = self.network_api.allocate_for_instance( [ 772.961359] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 772.961359] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] created_port_ids = self._update_ports_for_instance( [ 772.961359] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 772.961359] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] with excutils.save_and_reraise_exception(): [ 772.961359] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 772.961359] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] self.force_reraise() [ 772.961359] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 772.961359] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] raise self.value [ 772.961359] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 772.961783] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] updated_port = self._update_port( [ 772.961783] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 772.961783] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] _ensure_no_port_binding_failure(port) [ 772.961783] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 772.961783] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] raise exception.PortBindingFailed(port_id=port['id']) [ 772.961783] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] nova.exception.PortBindingFailed: Binding failed for port 955316c4-ca43-4931-a1f2-cf43d26ffa46, please check neutron logs for more information. [ 772.961783] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] [ 772.961783] env[61906]: INFO nova.compute.manager [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Terminating instance [ 772.964114] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Acquiring lock "refresh_cache-a048328a-d4a8-420e-9cc8-047254d7d1bf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.964276] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Acquired lock "refresh_cache-a048328a-d4a8-420e-9cc8-047254d7d1bf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.964443] env[61906]: DEBUG nova.network.neutron [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 772.969816] env[61906]: DEBUG nova.network.neutron [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 773.069600] env[61906]: DEBUG nova.network.neutron [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.077145] env[61906]: DEBUG nova.network.neutron [-] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.386697] env[61906]: DEBUG oslo_concurrency.lockutils [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.386937] env[61906]: DEBUG oslo_concurrency.lockutils [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.484878] env[61906]: DEBUG nova.network.neutron [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 773.572799] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Releasing lock "refresh_cache-bf056a94-9e9b-4296-b621-7e8a7338330b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.573066] env[61906]: DEBUG nova.compute.manager [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 773.573266] env[61906]: DEBUG nova.compute.manager [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 773.573568] env[61906]: DEBUG nova.network.neutron [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 773.580806] env[61906]: INFO nova.compute.manager [-] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Took 1.03 seconds to deallocate network for instance. [ 773.586705] env[61906]: DEBUG nova.network.neutron [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.591020] env[61906]: DEBUG nova.compute.claims [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 773.591020] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.594618] env[61906]: DEBUG nova.network.neutron [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 773.767769] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ddb163d-0ef4-44d6-9ab2-8dab7c459709 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.776688] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16e7c191-14be-4bc1-856c-7ad97275b9c8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.806224] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0010218-9c59-4ec8-a0bf-60af4573ab5e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.813138] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f423fb7-bc93-4c98-bcee-1cd6ee81e88c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.825693] env[61906]: DEBUG nova.compute.provider_tree [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 773.976623] env[61906]: DEBUG nova.compute.manager [req-c17d2aae-9e0a-4d2f-b50e-e7759279af2e req-b7a7a415-3495-4ec4-8857-0978f4c9b64f service nova] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Received event network-changed-955316c4-ca43-4931-a1f2-cf43d26ffa46 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 773.976831] env[61906]: DEBUG nova.compute.manager [req-c17d2aae-9e0a-4d2f-b50e-e7759279af2e req-b7a7a415-3495-4ec4-8857-0978f4c9b64f service nova] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Refreshing instance network info cache due to event network-changed-955316c4-ca43-4931-a1f2-cf43d26ffa46. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 773.977041] env[61906]: DEBUG oslo_concurrency.lockutils [req-c17d2aae-9e0a-4d2f-b50e-e7759279af2e req-b7a7a415-3495-4ec4-8857-0978f4c9b64f service nova] Acquiring lock "refresh_cache-a048328a-d4a8-420e-9cc8-047254d7d1bf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.089344] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Releasing lock "refresh_cache-a048328a-d4a8-420e-9cc8-047254d7d1bf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.089794] env[61906]: DEBUG nova.compute.manager [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 774.090212] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 774.090296] env[61906]: DEBUG oslo_concurrency.lockutils [req-c17d2aae-9e0a-4d2f-b50e-e7759279af2e req-b7a7a415-3495-4ec4-8857-0978f4c9b64f service nova] Acquired lock "refresh_cache-a048328a-d4a8-420e-9cc8-047254d7d1bf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.090470] env[61906]: DEBUG nova.network.neutron [req-c17d2aae-9e0a-4d2f-b50e-e7759279af2e req-b7a7a415-3495-4ec4-8857-0978f4c9b64f service nova] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Refreshing network info cache for port 955316c4-ca43-4931-a1f2-cf43d26ffa46 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 774.091551] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8660f50a-aceb-431b-abea-2de2dd179095 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.096602] env[61906]: DEBUG nova.network.neutron [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.102594] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-076a4413-7915-4859-8092-a2c8938bcb5d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.128144] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a048328a-d4a8-420e-9cc8-047254d7d1bf could not be found. [ 774.128144] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 774.128144] env[61906]: INFO nova.compute.manager [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Took 0.04 seconds to destroy the instance on the hypervisor. [ 774.129063] env[61906]: DEBUG oslo.service.loopingcall [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 774.129063] env[61906]: DEBUG nova.compute.manager [-] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 774.129063] env[61906]: DEBUG nova.network.neutron [-] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 774.149141] env[61906]: DEBUG nova.network.neutron [-] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.330438] env[61906]: DEBUG nova.scheduler.client.report [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 774.600661] env[61906]: INFO nova.compute.manager [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] [instance: bf056a94-9e9b-4296-b621-7e8a7338330b] Took 1.03 seconds to deallocate network for instance. [ 774.611056] env[61906]: DEBUG nova.network.neutron [req-c17d2aae-9e0a-4d2f-b50e-e7759279af2e req-b7a7a415-3495-4ec4-8857-0978f4c9b64f service nova] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.650849] env[61906]: DEBUG nova.network.neutron [-] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.689615] env[61906]: DEBUG nova.network.neutron [req-c17d2aae-9e0a-4d2f-b50e-e7759279af2e req-b7a7a415-3495-4ec4-8857-0978f4c9b64f service nova] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.836345] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.395s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.837283] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.767s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.838738] env[61906]: INFO nova.compute.claims [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 775.153635] env[61906]: INFO nova.compute.manager [-] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Took 1.02 seconds to deallocate network for instance. [ 775.156503] env[61906]: DEBUG nova.compute.claims [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 775.156696] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.193032] env[61906]: DEBUG oslo_concurrency.lockutils [req-c17d2aae-9e0a-4d2f-b50e-e7759279af2e req-b7a7a415-3495-4ec4-8857-0978f4c9b64f service nova] Releasing lock "refresh_cache-a048328a-d4a8-420e-9cc8-047254d7d1bf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.194025] env[61906]: DEBUG nova.compute.manager [req-c17d2aae-9e0a-4d2f-b50e-e7759279af2e req-b7a7a415-3495-4ec4-8857-0978f4c9b64f service nova] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Received event network-vif-deleted-955316c4-ca43-4931-a1f2-cf43d26ffa46 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 775.342413] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Acquiring lock "dada9e91-832e-4525-82fd-6cb90a65ca10" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.342677] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Lock "dada9e91-832e-4525-82fd-6cb90a65ca10" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.628024] env[61906]: INFO nova.scheduler.client.report [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Deleted allocations for instance bf056a94-9e9b-4296-b621-7e8a7338330b [ 775.846880] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Lock "dada9e91-832e-4525-82fd-6cb90a65ca10" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.504s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.847505] env[61906]: DEBUG nova.compute.manager [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 776.115246] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b937f46-7250-4b8e-b521-8ed0cb44e5ff {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.122782] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-493f9db7-aa87-452e-9978-42c189d7cea2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.152481] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa01de75-7b2f-4dcd-bab0-7c72e25e12b3 tempest-ImagesNegativeTestJSON-138158426 tempest-ImagesNegativeTestJSON-138158426-project-member] Lock "bf056a94-9e9b-4296-b621-7e8a7338330b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 172.423s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.155047] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f5ac776-da35-49a5-8a3e-c97b442e4f06 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.162773] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04392d1b-8383-4161-b046-104f4b533f61 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.177056] env[61906]: DEBUG nova.compute.provider_tree [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.352818] env[61906]: DEBUG nova.compute.utils [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 776.354584] env[61906]: DEBUG nova.compute.manager [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 776.354757] env[61906]: DEBUG nova.network.neutron [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 776.399104] env[61906]: DEBUG nova.policy [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '211680a6d83e4b19ba6093d5a9fc0a49', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '04f51e380fe44e4e8a583ded37d2d4d8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 776.658241] env[61906]: DEBUG nova.compute.manager [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 776.680130] env[61906]: DEBUG nova.scheduler.client.report [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 776.725690] env[61906]: DEBUG nova.network.neutron [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Successfully created port: 10d9cc70-a781-45a3-ba03-c4d9ed3c6d27 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 776.857963] env[61906]: DEBUG nova.compute.manager [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 777.179955] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.190263] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.350s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.190263] env[61906]: DEBUG nova.compute.manager [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 777.191263] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.097s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.192834] env[61906]: INFO nova.compute.claims [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 777.577336] env[61906]: DEBUG nova.compute.manager [req-b34cffcd-d7f0-49db-bbbe-50894c3987cd req-0a1fe1ad-7de5-4e41-ae69-d4d28ffbb2f3 service nova] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Received event network-changed-10d9cc70-a781-45a3-ba03-c4d9ed3c6d27 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 777.577570] env[61906]: DEBUG nova.compute.manager [req-b34cffcd-d7f0-49db-bbbe-50894c3987cd req-0a1fe1ad-7de5-4e41-ae69-d4d28ffbb2f3 service nova] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Refreshing instance network info cache due to event network-changed-10d9cc70-a781-45a3-ba03-c4d9ed3c6d27. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 777.577723] env[61906]: DEBUG oslo_concurrency.lockutils [req-b34cffcd-d7f0-49db-bbbe-50894c3987cd req-0a1fe1ad-7de5-4e41-ae69-d4d28ffbb2f3 service nova] Acquiring lock "refresh_cache-1b289a3d-7958-492b-ab9d-4ba8d3ffdae8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.577873] env[61906]: DEBUG oslo_concurrency.lockutils [req-b34cffcd-d7f0-49db-bbbe-50894c3987cd req-0a1fe1ad-7de5-4e41-ae69-d4d28ffbb2f3 service nova] Acquired lock "refresh_cache-1b289a3d-7958-492b-ab9d-4ba8d3ffdae8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.578828] env[61906]: DEBUG nova.network.neutron [req-b34cffcd-d7f0-49db-bbbe-50894c3987cd req-0a1fe1ad-7de5-4e41-ae69-d4d28ffbb2f3 service nova] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Refreshing network info cache for port 10d9cc70-a781-45a3-ba03-c4d9ed3c6d27 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 777.698762] env[61906]: DEBUG nova.compute.utils [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 777.703263] env[61906]: DEBUG nova.compute.manager [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 777.703263] env[61906]: DEBUG nova.network.neutron [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 777.794754] env[61906]: DEBUG nova.policy [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '957f22c5d01342a098974e22a4565c6f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f697463fb3a64ab1b5e66465f8d2603d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 777.842418] env[61906]: ERROR nova.compute.manager [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 10d9cc70-a781-45a3-ba03-c4d9ed3c6d27, please check neutron logs for more information. [ 777.842418] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 777.842418] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 777.842418] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 777.842418] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 777.842418] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 777.842418] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 777.842418] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 777.842418] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 777.842418] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 777.842418] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 777.842418] env[61906]: ERROR nova.compute.manager raise self.value [ 777.842418] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 777.842418] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 777.842418] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 777.842418] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 777.843089] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 777.843089] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 777.843089] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 10d9cc70-a781-45a3-ba03-c4d9ed3c6d27, please check neutron logs for more information. [ 777.843089] env[61906]: ERROR nova.compute.manager [ 777.843089] env[61906]: Traceback (most recent call last): [ 777.843089] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 777.843089] env[61906]: listener.cb(fileno) [ 777.843089] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 777.843089] env[61906]: result = function(*args, **kwargs) [ 777.843089] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 777.843089] env[61906]: return func(*args, **kwargs) [ 777.843089] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 777.843089] env[61906]: raise e [ 777.843089] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 777.843089] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 777.843089] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 777.843089] env[61906]: created_port_ids = self._update_ports_for_instance( [ 777.843089] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 777.843089] env[61906]: with excutils.save_and_reraise_exception(): [ 777.843089] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 777.843089] env[61906]: self.force_reraise() [ 777.843089] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 777.843089] env[61906]: raise self.value [ 777.843089] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 777.843089] env[61906]: updated_port = self._update_port( [ 777.843089] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 777.843089] env[61906]: _ensure_no_port_binding_failure(port) [ 777.843089] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 777.843089] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 777.844308] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 10d9cc70-a781-45a3-ba03-c4d9ed3c6d27, please check neutron logs for more information. [ 777.844308] env[61906]: Removing descriptor: 18 [ 777.872789] env[61906]: DEBUG nova.compute.manager [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 777.933011] env[61906]: DEBUG nova.virt.hardware [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 777.933266] env[61906]: DEBUG nova.virt.hardware [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 777.933423] env[61906]: DEBUG nova.virt.hardware [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 777.933604] env[61906]: DEBUG nova.virt.hardware [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 777.933748] env[61906]: DEBUG nova.virt.hardware [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 777.933893] env[61906]: DEBUG nova.virt.hardware [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 777.934117] env[61906]: DEBUG nova.virt.hardware [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 777.934279] env[61906]: DEBUG nova.virt.hardware [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 777.934449] env[61906]: DEBUG nova.virt.hardware [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 777.934607] env[61906]: DEBUG nova.virt.hardware [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 777.934773] env[61906]: DEBUG nova.virt.hardware [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 777.935703] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e054513f-dbb2-4c52-85cc-fb03d4b83091 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.944313] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7184596a-0ece-47e1-9be5-b6d996437a3e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.958592] env[61906]: ERROR nova.compute.manager [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 10d9cc70-a781-45a3-ba03-c4d9ed3c6d27, please check neutron logs for more information. [ 777.958592] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Traceback (most recent call last): [ 777.958592] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 777.958592] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] yield resources [ 777.958592] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 777.958592] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] self.driver.spawn(context, instance, image_meta, [ 777.958592] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 777.958592] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 777.958592] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 777.958592] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] vm_ref = self.build_virtual_machine(instance, [ 777.958592] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 777.959306] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] vif_infos = vmwarevif.get_vif_info(self._session, [ 777.959306] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 777.959306] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] for vif in network_info: [ 777.959306] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 777.959306] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] return self._sync_wrapper(fn, *args, **kwargs) [ 777.959306] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 777.959306] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] self.wait() [ 777.959306] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 777.959306] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] self[:] = self._gt.wait() [ 777.959306] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 777.959306] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] return self._exit_event.wait() [ 777.959306] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 777.959306] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] current.throw(*self._exc) [ 777.959967] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 777.959967] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] result = function(*args, **kwargs) [ 777.959967] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 777.959967] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] return func(*args, **kwargs) [ 777.959967] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 777.959967] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] raise e [ 777.959967] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 777.959967] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] nwinfo = self.network_api.allocate_for_instance( [ 777.959967] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 777.959967] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] created_port_ids = self._update_ports_for_instance( [ 777.959967] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 777.959967] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] with excutils.save_and_reraise_exception(): [ 777.959967] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 777.960503] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] self.force_reraise() [ 777.960503] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 777.960503] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] raise self.value [ 777.960503] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 777.960503] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] updated_port = self._update_port( [ 777.960503] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 777.960503] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] _ensure_no_port_binding_failure(port) [ 777.960503] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 777.960503] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] raise exception.PortBindingFailed(port_id=port['id']) [ 777.960503] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] nova.exception.PortBindingFailed: Binding failed for port 10d9cc70-a781-45a3-ba03-c4d9ed3c6d27, please check neutron logs for more information. [ 777.960503] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] [ 777.960503] env[61906]: INFO nova.compute.manager [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Terminating instance [ 777.961260] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Acquiring lock "refresh_cache-1b289a3d-7958-492b-ab9d-4ba8d3ffdae8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.160847] env[61906]: DEBUG nova.network.neutron [req-b34cffcd-d7f0-49db-bbbe-50894c3987cd req-0a1fe1ad-7de5-4e41-ae69-d4d28ffbb2f3 service nova] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.211425] env[61906]: DEBUG nova.compute.manager [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 778.309392] env[61906]: DEBUG nova.network.neutron [req-b34cffcd-d7f0-49db-bbbe-50894c3987cd req-0a1fe1ad-7de5-4e41-ae69-d4d28ffbb2f3 service nova] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.499410] env[61906]: DEBUG nova.network.neutron [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Successfully created port: 558793cc-c0df-4085-9518-c5b51652e2c5 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 778.554365] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ad4470f-a911-43e8-ad80-45d78bd32325 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.561345] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f936a903-089b-4b6b-ae82-feb7edc961ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.591276] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2c9cfb4-24c1-4a5f-8355-a5635efdeb9d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.598433] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa86ff9f-da5a-4296-8094-e2286b03764c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.616721] env[61906]: DEBUG nova.compute.provider_tree [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.812345] env[61906]: DEBUG oslo_concurrency.lockutils [req-b34cffcd-d7f0-49db-bbbe-50894c3987cd req-0a1fe1ad-7de5-4e41-ae69-d4d28ffbb2f3 service nova] Releasing lock "refresh_cache-1b289a3d-7958-492b-ab9d-4ba8d3ffdae8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.812762] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Acquired lock "refresh_cache-1b289a3d-7958-492b-ab9d-4ba8d3ffdae8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.812955] env[61906]: DEBUG nova.network.neutron [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 779.119614] env[61906]: DEBUG nova.scheduler.client.report [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 779.220697] env[61906]: DEBUG nova.compute.manager [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 779.246010] env[61906]: DEBUG nova.virt.hardware [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 779.246262] env[61906]: DEBUG nova.virt.hardware [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 779.246420] env[61906]: DEBUG nova.virt.hardware [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 779.246600] env[61906]: DEBUG nova.virt.hardware [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 779.246742] env[61906]: DEBUG nova.virt.hardware [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 779.246886] env[61906]: DEBUG nova.virt.hardware [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 779.247103] env[61906]: DEBUG nova.virt.hardware [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 779.247265] env[61906]: DEBUG nova.virt.hardware [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 779.247427] env[61906]: DEBUG nova.virt.hardware [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 779.247586] env[61906]: DEBUG nova.virt.hardware [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 779.247749] env[61906]: DEBUG nova.virt.hardware [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 779.248603] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e708b666-810d-4b3a-9caa-1d977171b241 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.256277] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4819c704-53f6-410d-ba2c-4c5881fec86b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.343123] env[61906]: DEBUG nova.network.neutron [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 779.492720] env[61906]: DEBUG nova.network.neutron [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.519935] env[61906]: ERROR nova.compute.manager [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 558793cc-c0df-4085-9518-c5b51652e2c5, please check neutron logs for more information. [ 779.519935] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 779.519935] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 779.519935] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 779.519935] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 779.519935] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 779.519935] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 779.519935] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 779.519935] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.519935] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 779.519935] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.519935] env[61906]: ERROR nova.compute.manager raise self.value [ 779.519935] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 779.519935] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 779.519935] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.519935] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 779.520576] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.520576] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 779.520576] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 558793cc-c0df-4085-9518-c5b51652e2c5, please check neutron logs for more information. [ 779.520576] env[61906]: ERROR nova.compute.manager [ 779.520576] env[61906]: Traceback (most recent call last): [ 779.520576] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 779.520576] env[61906]: listener.cb(fileno) [ 779.520576] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 779.520576] env[61906]: result = function(*args, **kwargs) [ 779.520576] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 779.520576] env[61906]: return func(*args, **kwargs) [ 779.520576] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 779.520576] env[61906]: raise e [ 779.520576] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 779.520576] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 779.520576] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 779.520576] env[61906]: created_port_ids = self._update_ports_for_instance( [ 779.520576] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 779.520576] env[61906]: with excutils.save_and_reraise_exception(): [ 779.520576] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.520576] env[61906]: self.force_reraise() [ 779.520576] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.520576] env[61906]: raise self.value [ 779.520576] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 779.520576] env[61906]: updated_port = self._update_port( [ 779.520576] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.520576] env[61906]: _ensure_no_port_binding_failure(port) [ 779.520576] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.520576] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 779.521656] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 558793cc-c0df-4085-9518-c5b51652e2c5, please check neutron logs for more information. [ 779.521656] env[61906]: Removing descriptor: 15 [ 779.521656] env[61906]: ERROR nova.compute.manager [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 558793cc-c0df-4085-9518-c5b51652e2c5, please check neutron logs for more information. [ 779.521656] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Traceback (most recent call last): [ 779.521656] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 779.521656] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] yield resources [ 779.521656] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 779.521656] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] self.driver.spawn(context, instance, image_meta, [ 779.521656] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 779.521656] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] self._vmops.spawn(context, instance, image_meta, injected_files, [ 779.521656] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 779.521656] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] vm_ref = self.build_virtual_machine(instance, [ 779.522068] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 779.522068] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] vif_infos = vmwarevif.get_vif_info(self._session, [ 779.522068] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 779.522068] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] for vif in network_info: [ 779.522068] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 779.522068] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] return self._sync_wrapper(fn, *args, **kwargs) [ 779.522068] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 779.522068] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] self.wait() [ 779.522068] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 779.522068] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] self[:] = self._gt.wait() [ 779.522068] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 779.522068] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] return self._exit_event.wait() [ 779.522068] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 779.522548] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] result = hub.switch() [ 779.522548] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 779.522548] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] return self.greenlet.switch() [ 779.522548] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 779.522548] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] result = function(*args, **kwargs) [ 779.522548] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 779.522548] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] return func(*args, **kwargs) [ 779.522548] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 779.522548] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] raise e [ 779.522548] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 779.522548] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] nwinfo = self.network_api.allocate_for_instance( [ 779.522548] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 779.522548] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] created_port_ids = self._update_ports_for_instance( [ 779.523039] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 779.523039] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] with excutils.save_and_reraise_exception(): [ 779.523039] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.523039] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] self.force_reraise() [ 779.523039] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.523039] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] raise self.value [ 779.523039] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 779.523039] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] updated_port = self._update_port( [ 779.523039] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.523039] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] _ensure_no_port_binding_failure(port) [ 779.523039] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.523039] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] raise exception.PortBindingFailed(port_id=port['id']) [ 779.523485] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] nova.exception.PortBindingFailed: Binding failed for port 558793cc-c0df-4085-9518-c5b51652e2c5, please check neutron logs for more information. [ 779.523485] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] [ 779.523485] env[61906]: INFO nova.compute.manager [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Terminating instance [ 779.523485] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Acquiring lock "refresh_cache-830e7d50-5870-4e1f-8f2f-26840ba06d33" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.523485] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Acquired lock "refresh_cache-830e7d50-5870-4e1f-8f2f-26840ba06d33" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.523485] env[61906]: DEBUG nova.network.neutron [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 779.607486] env[61906]: DEBUG nova.compute.manager [req-1962e130-55b9-4603-b15f-6661260b832c req-4450fb45-4830-47bb-943c-3b7c9f1199dd service nova] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Received event network-vif-deleted-10d9cc70-a781-45a3-ba03-c4d9ed3c6d27 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 779.607588] env[61906]: DEBUG nova.compute.manager [req-1962e130-55b9-4603-b15f-6661260b832c req-4450fb45-4830-47bb-943c-3b7c9f1199dd service nova] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Received event network-changed-558793cc-c0df-4085-9518-c5b51652e2c5 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 779.607725] env[61906]: DEBUG nova.compute.manager [req-1962e130-55b9-4603-b15f-6661260b832c req-4450fb45-4830-47bb-943c-3b7c9f1199dd service nova] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Refreshing instance network info cache due to event network-changed-558793cc-c0df-4085-9518-c5b51652e2c5. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 779.607911] env[61906]: DEBUG oslo_concurrency.lockutils [req-1962e130-55b9-4603-b15f-6661260b832c req-4450fb45-4830-47bb-943c-3b7c9f1199dd service nova] Acquiring lock "refresh_cache-830e7d50-5870-4e1f-8f2f-26840ba06d33" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.624974] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.434s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.625476] env[61906]: DEBUG nova.compute.manager [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 779.627815] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.749s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.997298] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Releasing lock "refresh_cache-1b289a3d-7958-492b-ab9d-4ba8d3ffdae8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.997733] env[61906]: DEBUG nova.compute.manager [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 779.997934] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 779.998862] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bae22e1c-1bfc-47ee-93fc-f6f1383a1d5a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.007487] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3960f6c-6fa4-47fb-9c1c-91cafd48d7d1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.029745] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8 could not be found. [ 780.029946] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 780.030137] env[61906]: INFO nova.compute.manager [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Took 0.03 seconds to destroy the instance on the hypervisor. [ 780.030375] env[61906]: DEBUG oslo.service.loopingcall [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 780.030594] env[61906]: DEBUG nova.compute.manager [-] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 780.030688] env[61906]: DEBUG nova.network.neutron [-] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 780.053934] env[61906]: DEBUG nova.network.neutron [-] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.055587] env[61906]: DEBUG nova.network.neutron [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.106148] env[61906]: DEBUG nova.network.neutron [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.132349] env[61906]: DEBUG nova.compute.utils [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 780.136091] env[61906]: DEBUG nova.compute.manager [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 780.136269] env[61906]: DEBUG nova.network.neutron [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 780.183696] env[61906]: DEBUG nova.policy [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bf22a6ff9c8541b59ba9a72529c48027', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5972cff43b0b4d5ab21ed30f126d94ef', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 780.430287] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-068a6c9a-ad57-4a4b-883e-1e18e6792262 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.439246] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e81bb700-2c6d-4a31-8c35-1efbd3f7346b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.471689] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8997262-f502-44ef-8cdb-4171c3b14ee7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.479167] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72a8f171-367f-41e2-82b3-e8fb267e1ddf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.492313] env[61906]: DEBUG nova.compute.provider_tree [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.497783] env[61906]: DEBUG nova.network.neutron [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Successfully created port: 5d3b6ed6-477e-4457-94a1-2b812721d3e3 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 780.558374] env[61906]: DEBUG nova.network.neutron [-] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.608555] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Releasing lock "refresh_cache-830e7d50-5870-4e1f-8f2f-26840ba06d33" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.609016] env[61906]: DEBUG nova.compute.manager [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 780.609398] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 780.609737] env[61906]: DEBUG oslo_concurrency.lockutils [req-1962e130-55b9-4603-b15f-6661260b832c req-4450fb45-4830-47bb-943c-3b7c9f1199dd service nova] Acquired lock "refresh_cache-830e7d50-5870-4e1f-8f2f-26840ba06d33" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.609919] env[61906]: DEBUG nova.network.neutron [req-1962e130-55b9-4603-b15f-6661260b832c req-4450fb45-4830-47bb-943c-3b7c9f1199dd service nova] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Refreshing network info cache for port 558793cc-c0df-4085-9518-c5b51652e2c5 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 780.610957] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-095e1a9a-4376-4a1e-bb5e-2b3943b7f3f5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.621645] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ba21b64-6ee6-49fc-904b-a4ac44e624ec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.643566] env[61906]: DEBUG nova.compute.manager [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 780.646236] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 830e7d50-5870-4e1f-8f2f-26840ba06d33 could not be found. [ 780.646439] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 780.646611] env[61906]: INFO nova.compute.manager [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Took 0.04 seconds to destroy the instance on the hypervisor. [ 780.646843] env[61906]: DEBUG oslo.service.loopingcall [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 780.647413] env[61906]: DEBUG nova.compute.manager [-] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 780.647504] env[61906]: DEBUG nova.network.neutron [-] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 780.677954] env[61906]: DEBUG nova.network.neutron [-] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.998261] env[61906]: DEBUG nova.scheduler.client.report [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 781.061253] env[61906]: INFO nova.compute.manager [-] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Took 1.03 seconds to deallocate network for instance. [ 781.063602] env[61906]: DEBUG nova.compute.claims [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 781.063782] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.148214] env[61906]: DEBUG nova.network.neutron [req-1962e130-55b9-4603-b15f-6661260b832c req-4450fb45-4830-47bb-943c-3b7c9f1199dd service nova] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 781.181261] env[61906]: DEBUG nova.network.neutron [-] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.219235] env[61906]: DEBUG nova.network.neutron [req-1962e130-55b9-4603-b15f-6661260b832c req-4450fb45-4830-47bb-943c-3b7c9f1199dd service nova] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.504781] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.877s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.505435] env[61906]: ERROR nova.compute.manager [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a41cd4c0-d7ec-48e3-9c57-a2d24241b560, please check neutron logs for more information. [ 781.505435] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Traceback (most recent call last): [ 781.505435] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 781.505435] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] self.driver.spawn(context, instance, image_meta, [ 781.505435] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 781.505435] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] self._vmops.spawn(context, instance, image_meta, injected_files, [ 781.505435] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 781.505435] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] vm_ref = self.build_virtual_machine(instance, [ 781.505435] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 781.505435] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] vif_infos = vmwarevif.get_vif_info(self._session, [ 781.505435] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 781.505815] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] for vif in network_info: [ 781.505815] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 781.505815] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] return self._sync_wrapper(fn, *args, **kwargs) [ 781.505815] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 781.505815] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] self.wait() [ 781.505815] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 781.505815] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] self[:] = self._gt.wait() [ 781.505815] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 781.505815] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] return self._exit_event.wait() [ 781.505815] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 781.505815] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] current.throw(*self._exc) [ 781.505815] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.505815] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] result = function(*args, **kwargs) [ 781.506233] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 781.506233] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] return func(*args, **kwargs) [ 781.506233] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 781.506233] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] raise e [ 781.506233] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.506233] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] nwinfo = self.network_api.allocate_for_instance( [ 781.506233] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.506233] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] created_port_ids = self._update_ports_for_instance( [ 781.506233] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.506233] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] with excutils.save_and_reraise_exception(): [ 781.506233] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.506233] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] self.force_reraise() [ 781.506233] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.506629] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] raise self.value [ 781.506629] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.506629] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] updated_port = self._update_port( [ 781.506629] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.506629] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] _ensure_no_port_binding_failure(port) [ 781.506629] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.506629] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] raise exception.PortBindingFailed(port_id=port['id']) [ 781.506629] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] nova.exception.PortBindingFailed: Binding failed for port a41cd4c0-d7ec-48e3-9c57-a2d24241b560, please check neutron logs for more information. [ 781.506629] env[61906]: ERROR nova.compute.manager [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] [ 781.506629] env[61906]: DEBUG nova.compute.utils [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Binding failed for port a41cd4c0-d7ec-48e3-9c57-a2d24241b560, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 781.507444] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.898s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.510301] env[61906]: DEBUG nova.compute.manager [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Build of instance b071a9b5-3416-4898-9ae2-ef62186c1843 was re-scheduled: Binding failed for port a41cd4c0-d7ec-48e3-9c57-a2d24241b560, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 781.510720] env[61906]: DEBUG nova.compute.manager [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 781.510946] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Acquiring lock "refresh_cache-b071a9b5-3416-4898-9ae2-ef62186c1843" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.514039] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Acquired lock "refresh_cache-b071a9b5-3416-4898-9ae2-ef62186c1843" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.515892] env[61906]: DEBUG nova.network.neutron [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 781.534857] env[61906]: ERROR nova.compute.manager [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5d3b6ed6-477e-4457-94a1-2b812721d3e3, please check neutron logs for more information. [ 781.534857] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 781.534857] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.534857] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 781.534857] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.534857] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 781.534857] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.534857] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 781.534857] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.534857] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 781.534857] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.534857] env[61906]: ERROR nova.compute.manager raise self.value [ 781.534857] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.534857] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 781.534857] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.534857] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 781.535607] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.535607] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 781.535607] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5d3b6ed6-477e-4457-94a1-2b812721d3e3, please check neutron logs for more information. [ 781.535607] env[61906]: ERROR nova.compute.manager [ 781.535607] env[61906]: Traceback (most recent call last): [ 781.535607] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 781.535607] env[61906]: listener.cb(fileno) [ 781.535607] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.535607] env[61906]: result = function(*args, **kwargs) [ 781.535607] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 781.535607] env[61906]: return func(*args, **kwargs) [ 781.535607] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 781.535607] env[61906]: raise e [ 781.535607] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.535607] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 781.535607] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.535607] env[61906]: created_port_ids = self._update_ports_for_instance( [ 781.535607] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.535607] env[61906]: with excutils.save_and_reraise_exception(): [ 781.535607] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.535607] env[61906]: self.force_reraise() [ 781.535607] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.535607] env[61906]: raise self.value [ 781.535607] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.535607] env[61906]: updated_port = self._update_port( [ 781.535607] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.535607] env[61906]: _ensure_no_port_binding_failure(port) [ 781.535607] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.535607] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 781.536637] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 5d3b6ed6-477e-4457-94a1-2b812721d3e3, please check neutron logs for more information. [ 781.536637] env[61906]: Removing descriptor: 15 [ 781.631618] env[61906]: DEBUG nova.compute.manager [req-d389fef7-daa7-4ae5-8768-5d3041bb846a req-45952e9f-54f6-4a36-887c-6f9967c44e6d service nova] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Received event network-changed-5d3b6ed6-477e-4457-94a1-2b812721d3e3 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 781.631674] env[61906]: DEBUG nova.compute.manager [req-d389fef7-daa7-4ae5-8768-5d3041bb846a req-45952e9f-54f6-4a36-887c-6f9967c44e6d service nova] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Refreshing instance network info cache due to event network-changed-5d3b6ed6-477e-4457-94a1-2b812721d3e3. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 781.631872] env[61906]: DEBUG oslo_concurrency.lockutils [req-d389fef7-daa7-4ae5-8768-5d3041bb846a req-45952e9f-54f6-4a36-887c-6f9967c44e6d service nova] Acquiring lock "refresh_cache-5ab321ec-655b-4f68-91aa-9b3db8c692f4" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.632044] env[61906]: DEBUG oslo_concurrency.lockutils [req-d389fef7-daa7-4ae5-8768-5d3041bb846a req-45952e9f-54f6-4a36-887c-6f9967c44e6d service nova] Acquired lock "refresh_cache-5ab321ec-655b-4f68-91aa-9b3db8c692f4" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.632237] env[61906]: DEBUG nova.network.neutron [req-d389fef7-daa7-4ae5-8768-5d3041bb846a req-45952e9f-54f6-4a36-887c-6f9967c44e6d service nova] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Refreshing network info cache for port 5d3b6ed6-477e-4457-94a1-2b812721d3e3 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 781.652716] env[61906]: DEBUG nova.compute.manager [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 781.677572] env[61906]: DEBUG nova.virt.hardware [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:21:34Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='75820dea-f9b9-4f26-815d-e6531c7a0c81',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-799327559',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 781.677805] env[61906]: DEBUG nova.virt.hardware [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 781.677963] env[61906]: DEBUG nova.virt.hardware [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 781.678157] env[61906]: DEBUG nova.virt.hardware [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 781.678304] env[61906]: DEBUG nova.virt.hardware [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 781.678449] env[61906]: DEBUG nova.virt.hardware [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 781.678649] env[61906]: DEBUG nova.virt.hardware [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 781.678804] env[61906]: DEBUG nova.virt.hardware [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 781.678963] env[61906]: DEBUG nova.virt.hardware [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 781.679143] env[61906]: DEBUG nova.virt.hardware [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 781.679410] env[61906]: DEBUG nova.virt.hardware [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 781.680459] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bb8f582-7f80-4bbe-8053-247a2857b135 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.683079] env[61906]: INFO nova.compute.manager [-] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Took 1.04 seconds to deallocate network for instance. [ 781.685203] env[61906]: DEBUG nova.compute.claims [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 781.685393] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.689904] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc3d045c-659f-4e00-8332-be06040a0e01 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.704494] env[61906]: ERROR nova.compute.manager [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5d3b6ed6-477e-4457-94a1-2b812721d3e3, please check neutron logs for more information. [ 781.704494] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Traceback (most recent call last): [ 781.704494] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 781.704494] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] yield resources [ 781.704494] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 781.704494] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] self.driver.spawn(context, instance, image_meta, [ 781.704494] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 781.704494] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 781.704494] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 781.704494] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] vm_ref = self.build_virtual_machine(instance, [ 781.704494] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 781.704864] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] vif_infos = vmwarevif.get_vif_info(self._session, [ 781.704864] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 781.704864] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] for vif in network_info: [ 781.704864] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 781.704864] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] return self._sync_wrapper(fn, *args, **kwargs) [ 781.704864] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 781.704864] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] self.wait() [ 781.704864] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 781.704864] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] self[:] = self._gt.wait() [ 781.704864] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 781.704864] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] return self._exit_event.wait() [ 781.704864] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 781.704864] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] current.throw(*self._exc) [ 781.706275] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.706275] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] result = function(*args, **kwargs) [ 781.706275] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 781.706275] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] return func(*args, **kwargs) [ 781.706275] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 781.706275] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] raise e [ 781.706275] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.706275] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] nwinfo = self.network_api.allocate_for_instance( [ 781.706275] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.706275] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] created_port_ids = self._update_ports_for_instance( [ 781.706275] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.706275] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] with excutils.save_and_reraise_exception(): [ 781.706275] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.706747] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] self.force_reraise() [ 781.706747] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.706747] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] raise self.value [ 781.706747] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.706747] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] updated_port = self._update_port( [ 781.706747] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.706747] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] _ensure_no_port_binding_failure(port) [ 781.706747] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.706747] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] raise exception.PortBindingFailed(port_id=port['id']) [ 781.706747] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] nova.exception.PortBindingFailed: Binding failed for port 5d3b6ed6-477e-4457-94a1-2b812721d3e3, please check neutron logs for more information. [ 781.706747] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] [ 781.706747] env[61906]: INFO nova.compute.manager [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Terminating instance [ 781.707203] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Acquiring lock "refresh_cache-5ab321ec-655b-4f68-91aa-9b3db8c692f4" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.722604] env[61906]: DEBUG oslo_concurrency.lockutils [req-1962e130-55b9-4603-b15f-6661260b832c req-4450fb45-4830-47bb-943c-3b7c9f1199dd service nova] Releasing lock "refresh_cache-830e7d50-5870-4e1f-8f2f-26840ba06d33" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.722867] env[61906]: DEBUG nova.compute.manager [req-1962e130-55b9-4603-b15f-6661260b832c req-4450fb45-4830-47bb-943c-3b7c9f1199dd service nova] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Received event network-vif-deleted-558793cc-c0df-4085-9518-c5b51652e2c5 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 782.037473] env[61906]: DEBUG nova.network.neutron [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.132803] env[61906]: DEBUG nova.network.neutron [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.154270] env[61906]: DEBUG nova.network.neutron [req-d389fef7-daa7-4ae5-8768-5d3041bb846a req-45952e9f-54f6-4a36-887c-6f9967c44e6d service nova] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.205179] env[61906]: DEBUG nova.network.neutron [req-d389fef7-daa7-4ae5-8768-5d3041bb846a req-45952e9f-54f6-4a36-887c-6f9967c44e6d service nova] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.296198] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04af5c2f-e81e-4c6b-9511-223616e61fc5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.304779] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01e7ee1d-d24e-4de3-8928-205988c9e20c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.333847] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eec69d40-c769-4b2a-9bcd-fab334e42002 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.341105] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76e5795d-cc33-4a6b-a9e6-8d4b0b603d5d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.354168] env[61906]: DEBUG nova.compute.provider_tree [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.637200] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Releasing lock "refresh_cache-b071a9b5-3416-4898-9ae2-ef62186c1843" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.637465] env[61906]: DEBUG nova.compute.manager [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 782.637652] env[61906]: DEBUG nova.compute.manager [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 782.637821] env[61906]: DEBUG nova.network.neutron [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 782.653337] env[61906]: DEBUG nova.network.neutron [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.706622] env[61906]: DEBUG oslo_concurrency.lockutils [req-d389fef7-daa7-4ae5-8768-5d3041bb846a req-45952e9f-54f6-4a36-887c-6f9967c44e6d service nova] Releasing lock "refresh_cache-5ab321ec-655b-4f68-91aa-9b3db8c692f4" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.706866] env[61906]: DEBUG nova.compute.manager [req-d389fef7-daa7-4ae5-8768-5d3041bb846a req-45952e9f-54f6-4a36-887c-6f9967c44e6d service nova] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Received event network-vif-deleted-5d3b6ed6-477e-4457-94a1-2b812721d3e3 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 782.707234] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Acquired lock "refresh_cache-5ab321ec-655b-4f68-91aa-9b3db8c692f4" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.707404] env[61906]: DEBUG nova.network.neutron [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 782.857253] env[61906]: DEBUG nova.scheduler.client.report [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 783.156076] env[61906]: DEBUG nova.network.neutron [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.230649] env[61906]: DEBUG nova.network.neutron [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 783.285534] env[61906]: DEBUG nova.network.neutron [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.361350] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.854s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.361996] env[61906]: ERROR nova.compute.manager [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 331a64dd-86e1-492c-aa13-2a0dd0b47a77, please check neutron logs for more information. [ 783.361996] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Traceback (most recent call last): [ 783.361996] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 783.361996] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] self.driver.spawn(context, instance, image_meta, [ 783.361996] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 783.361996] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] self._vmops.spawn(context, instance, image_meta, injected_files, [ 783.361996] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 783.361996] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] vm_ref = self.build_virtual_machine(instance, [ 783.361996] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 783.361996] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] vif_infos = vmwarevif.get_vif_info(self._session, [ 783.361996] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 783.362490] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] for vif in network_info: [ 783.362490] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 783.362490] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] return self._sync_wrapper(fn, *args, **kwargs) [ 783.362490] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 783.362490] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] self.wait() [ 783.362490] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 783.362490] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] self[:] = self._gt.wait() [ 783.362490] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 783.362490] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] return self._exit_event.wait() [ 783.362490] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 783.362490] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] result = hub.switch() [ 783.362490] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 783.362490] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] return self.greenlet.switch() [ 783.362944] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 783.362944] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] result = function(*args, **kwargs) [ 783.362944] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 783.362944] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] return func(*args, **kwargs) [ 783.362944] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 783.362944] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] raise e [ 783.362944] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 783.362944] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] nwinfo = self.network_api.allocate_for_instance( [ 783.362944] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 783.362944] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] created_port_ids = self._update_ports_for_instance( [ 783.362944] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 783.362944] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] with excutils.save_and_reraise_exception(): [ 783.362944] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 783.363413] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] self.force_reraise() [ 783.363413] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 783.363413] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] raise self.value [ 783.363413] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 783.363413] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] updated_port = self._update_port( [ 783.363413] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 783.363413] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] _ensure_no_port_binding_failure(port) [ 783.363413] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 783.363413] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] raise exception.PortBindingFailed(port_id=port['id']) [ 783.363413] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] nova.exception.PortBindingFailed: Binding failed for port 331a64dd-86e1-492c-aa13-2a0dd0b47a77, please check neutron logs for more information. [ 783.363413] env[61906]: ERROR nova.compute.manager [instance: a07e49a5-e20a-49df-9326-a2e819934d15] [ 783.363839] env[61906]: DEBUG nova.compute.utils [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Binding failed for port 331a64dd-86e1-492c-aa13-2a0dd0b47a77, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 783.363950] env[61906]: DEBUG oslo_concurrency.lockutils [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.366668] env[61906]: DEBUG nova.compute.manager [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Build of instance a07e49a5-e20a-49df-9326-a2e819934d15 was re-scheduled: Binding failed for port 331a64dd-86e1-492c-aa13-2a0dd0b47a77, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 783.367099] env[61906]: DEBUG nova.compute.manager [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 783.367323] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Acquiring lock "refresh_cache-a07e49a5-e20a-49df-9326-a2e819934d15" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.367529] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Acquired lock "refresh_cache-a07e49a5-e20a-49df-9326-a2e819934d15" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.367629] env[61906]: DEBUG nova.network.neutron [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 783.659026] env[61906]: INFO nova.compute.manager [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: b071a9b5-3416-4898-9ae2-ef62186c1843] Took 1.02 seconds to deallocate network for instance. [ 783.788639] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Releasing lock "refresh_cache-5ab321ec-655b-4f68-91aa-9b3db8c692f4" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.789647] env[61906]: DEBUG nova.compute.manager [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 783.789782] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 783.790159] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-417ec497-c3ee-468b-a935-029c605b4808 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.800813] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f262b63-4077-4be2-8e0a-956189cecda2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.820917] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5ab321ec-655b-4f68-91aa-9b3db8c692f4 could not be found. [ 783.821170] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 783.821363] env[61906]: INFO nova.compute.manager [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Took 0.03 seconds to destroy the instance on the hypervisor. [ 783.821617] env[61906]: DEBUG oslo.service.loopingcall [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 783.822672] env[61906]: DEBUG nova.compute.manager [-] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 783.822672] env[61906]: DEBUG nova.network.neutron [-] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 783.845116] env[61906]: DEBUG nova.network.neutron [-] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 783.895499] env[61906]: DEBUG nova.network.neutron [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 783.979920] env[61906]: DEBUG nova.network.neutron [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.126237] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a6a7d77-70ec-42cd-a00c-2e8a75dbde4e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.133876] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5db8338-3818-442b-8f0e-a7c8fa553aa1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.165446] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46d57c05-e434-42d5-b0bd-46526ea9deee {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.172745] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9193097b-42f9-4929-af87-b37c0a95b96a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.185727] env[61906]: DEBUG nova.compute.provider_tree [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 784.347359] env[61906]: DEBUG nova.network.neutron [-] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.485263] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Releasing lock "refresh_cache-a07e49a5-e20a-49df-9326-a2e819934d15" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.485263] env[61906]: DEBUG nova.compute.manager [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 784.485263] env[61906]: DEBUG nova.compute.manager [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 784.485552] env[61906]: DEBUG nova.network.neutron [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 784.500372] env[61906]: DEBUG nova.network.neutron [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 784.689018] env[61906]: DEBUG nova.scheduler.client.report [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 784.693128] env[61906]: INFO nova.scheduler.client.report [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Deleted allocations for instance b071a9b5-3416-4898-9ae2-ef62186c1843 [ 784.849728] env[61906]: INFO nova.compute.manager [-] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Took 1.03 seconds to deallocate network for instance. [ 784.852178] env[61906]: DEBUG nova.compute.claims [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 784.852367] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.005681] env[61906]: DEBUG nova.network.neutron [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.200202] env[61906]: DEBUG oslo_concurrency.lockutils [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.836s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.201437] env[61906]: ERROR nova.compute.manager [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 56526b59-0198-4803-b772-890468f44b15, please check neutron logs for more information. [ 785.201437] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Traceback (most recent call last): [ 785.201437] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 785.201437] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] self.driver.spawn(context, instance, image_meta, [ 785.201437] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 785.201437] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 785.201437] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 785.201437] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] vm_ref = self.build_virtual_machine(instance, [ 785.201437] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 785.201437] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] vif_infos = vmwarevif.get_vif_info(self._session, [ 785.201437] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 785.201859] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] for vif in network_info: [ 785.201859] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 785.201859] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] return self._sync_wrapper(fn, *args, **kwargs) [ 785.201859] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 785.201859] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] self.wait() [ 785.201859] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 785.201859] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] self[:] = self._gt.wait() [ 785.201859] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 785.201859] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] return self._exit_event.wait() [ 785.201859] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 785.201859] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] result = hub.switch() [ 785.201859] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 785.201859] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] return self.greenlet.switch() [ 785.202293] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 785.202293] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] result = function(*args, **kwargs) [ 785.202293] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 785.202293] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] return func(*args, **kwargs) [ 785.202293] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 785.202293] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] raise e [ 785.202293] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 785.202293] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] nwinfo = self.network_api.allocate_for_instance( [ 785.202293] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 785.202293] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] created_port_ids = self._update_ports_for_instance( [ 785.202293] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 785.202293] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] with excutils.save_and_reraise_exception(): [ 785.202293] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 785.202683] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] self.force_reraise() [ 785.202683] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 785.202683] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] raise self.value [ 785.202683] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 785.202683] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] updated_port = self._update_port( [ 785.202683] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 785.202683] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] _ensure_no_port_binding_failure(port) [ 785.202683] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 785.202683] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] raise exception.PortBindingFailed(port_id=port['id']) [ 785.202683] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] nova.exception.PortBindingFailed: Binding failed for port 56526b59-0198-4803-b772-890468f44b15, please check neutron logs for more information. [ 785.202683] env[61906]: ERROR nova.compute.manager [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] [ 785.203034] env[61906]: DEBUG nova.compute.utils [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Binding failed for port 56526b59-0198-4803-b772-890468f44b15, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 785.203989] env[61906]: DEBUG nova.compute.manager [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Build of instance 58b454d7-5370-499a-93dc-8613ee997b4d was re-scheduled: Binding failed for port 56526b59-0198-4803-b772-890468f44b15, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 785.204419] env[61906]: DEBUG nova.compute.manager [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 785.204889] env[61906]: DEBUG oslo_concurrency.lockutils [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Acquiring lock "refresh_cache-58b454d7-5370-499a-93dc-8613ee997b4d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.205140] env[61906]: DEBUG oslo_concurrency.lockutils [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Acquired lock "refresh_cache-58b454d7-5370-499a-93dc-8613ee997b4d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.205338] env[61906]: DEBUG nova.network.neutron [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 785.206819] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7b9f33d6-4f8f-4f0a-a48e-b75f337d05cb tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Lock "b071a9b5-3416-4898-9ae2-ef62186c1843" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 180.347s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.207024] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.799s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.208671] env[61906]: INFO nova.compute.claims [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 785.509068] env[61906]: INFO nova.compute.manager [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] [instance: a07e49a5-e20a-49df-9326-a2e819934d15] Took 1.02 seconds to deallocate network for instance. [ 785.714262] env[61906]: DEBUG nova.compute.manager [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 785.723561] env[61906]: DEBUG nova.network.neutron [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 785.818734] env[61906]: DEBUG nova.network.neutron [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.238175] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.322572] env[61906]: DEBUG oslo_concurrency.lockutils [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Releasing lock "refresh_cache-58b454d7-5370-499a-93dc-8613ee997b4d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.322683] env[61906]: DEBUG nova.compute.manager [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 786.322799] env[61906]: DEBUG nova.compute.manager [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 786.323017] env[61906]: DEBUG nova.network.neutron [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 786.349014] env[61906]: DEBUG nova.network.neutron [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 786.486299] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22ac881-52fb-44ca-aa97-ec9f2e421e43 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.493998] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51726d12-b17e-4e62-9778-5dc3dbfb4b87 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.525961] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c046928b-29e0-4f9d-9d5d-111f07ebfa23 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.532297] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8488c0ba-7696-4fea-8229-0ca8f3c54250 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.544927] env[61906]: DEBUG nova.compute.provider_tree [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 786.546788] env[61906]: INFO nova.scheduler.client.report [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Deleted allocations for instance a07e49a5-e20a-49df-9326-a2e819934d15 [ 786.853944] env[61906]: DEBUG nova.network.neutron [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.053615] env[61906]: DEBUG nova.scheduler.client.report [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 787.056646] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7597469d-ad3e-4ad0-b551-38c5470d738a tempest-ListServerFiltersTestJSON-723894227 tempest-ListServerFiltersTestJSON-723894227-project-member] Lock "a07e49a5-e20a-49df-9326-a2e819934d15" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 181.146s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.357245] env[61906]: INFO nova.compute.manager [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] [instance: 58b454d7-5370-499a-93dc-8613ee997b4d] Took 1.03 seconds to deallocate network for instance. [ 787.559026] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.351s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.559026] env[61906]: DEBUG nova.compute.manager [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 787.561489] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 14.988s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.563270] env[61906]: DEBUG nova.compute.manager [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 788.072344] env[61906]: DEBUG nova.compute.utils [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 788.076250] env[61906]: DEBUG nova.compute.manager [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 788.076250] env[61906]: DEBUG nova.network.neutron [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 788.099180] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.154201] env[61906]: DEBUG nova.policy [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e40725b4b5043f09d5ea8401a6697d1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b22155abe44c460e8eac297a8a460486', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 788.391207] env[61906]: INFO nova.scheduler.client.report [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Deleted allocations for instance 58b454d7-5370-499a-93dc-8613ee997b4d [ 788.684174] env[61906]: DEBUG nova.compute.manager [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 788.709213] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 3df44b70-ee72-40d7-9129-b5e94e121b0a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 788.709427] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance a048328a-d4a8-420e-9cc8-047254d7d1bf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 788.709550] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 788.709662] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 830e7d50-5870-4e1f-8f2f-26840ba06d33 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 788.709771] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 5ab321ec-655b-4f68-91aa-9b3db8c692f4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 788.709878] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 7ad136a2-d850-4b68-af0f-73badf65426c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 788.747117] env[61906]: DEBUG nova.network.neutron [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Successfully created port: b01c48b8-7698-4df9-8653-500e06a0f61a {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 788.900610] env[61906]: DEBUG oslo_concurrency.lockutils [None req-20891306-b37b-4d79-be29-4927459c1b2d tempest-InstanceActionsNegativeTestJSON-70011383 tempest-InstanceActionsNegativeTestJSON-70011383-project-member] Lock "58b454d7-5370-499a-93dc-8613ee997b4d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 182.464s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.215636] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 7e452659-0a5e-4740-b7ed-1fd7b1917dac has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 789.403960] env[61906]: DEBUG nova.compute.manager [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 789.695450] env[61906]: DEBUG nova.compute.manager [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 789.719339] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance b935c7ba-7716-443b-b5da-9ee03cef793a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 789.722401] env[61906]: DEBUG nova.virt.hardware [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 789.722619] env[61906]: DEBUG nova.virt.hardware [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 789.722787] env[61906]: DEBUG nova.virt.hardware [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 789.722966] env[61906]: DEBUG nova.virt.hardware [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 789.723122] env[61906]: DEBUG nova.virt.hardware [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 789.723269] env[61906]: DEBUG nova.virt.hardware [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 789.723467] env[61906]: DEBUG nova.virt.hardware [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 789.723641] env[61906]: DEBUG nova.virt.hardware [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 789.723867] env[61906]: DEBUG nova.virt.hardware [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 789.728079] env[61906]: DEBUG nova.virt.hardware [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 789.728320] env[61906]: DEBUG nova.virt.hardware [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 789.729180] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d445e5-9766-4973-a3eb-4857100a9067 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.742760] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da352313-13af-4464-a423-0fbc9acdaf3f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.827875] env[61906]: DEBUG nova.compute.manager [req-4a46a711-41dd-4aaa-a314-76f1278d0bc7 req-b40ee905-5b06-476b-80a6-870a9be1fb20 service nova] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Received event network-changed-b01c48b8-7698-4df9-8653-500e06a0f61a {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 789.828121] env[61906]: DEBUG nova.compute.manager [req-4a46a711-41dd-4aaa-a314-76f1278d0bc7 req-b40ee905-5b06-476b-80a6-870a9be1fb20 service nova] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Refreshing instance network info cache due to event network-changed-b01c48b8-7698-4df9-8653-500e06a0f61a. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 789.828440] env[61906]: DEBUG oslo_concurrency.lockutils [req-4a46a711-41dd-4aaa-a314-76f1278d0bc7 req-b40ee905-5b06-476b-80a6-870a9be1fb20 service nova] Acquiring lock "refresh_cache-7ad136a2-d850-4b68-af0f-73badf65426c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.828542] env[61906]: DEBUG oslo_concurrency.lockutils [req-4a46a711-41dd-4aaa-a314-76f1278d0bc7 req-b40ee905-5b06-476b-80a6-870a9be1fb20 service nova] Acquired lock "refresh_cache-7ad136a2-d850-4b68-af0f-73badf65426c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.828734] env[61906]: DEBUG nova.network.neutron [req-4a46a711-41dd-4aaa-a314-76f1278d0bc7 req-b40ee905-5b06-476b-80a6-870a9be1fb20 service nova] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Refreshing network info cache for port b01c48b8-7698-4df9-8653-500e06a0f61a {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 789.933104] env[61906]: DEBUG oslo_concurrency.lockutils [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.032852] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Acquiring lock "3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.033101] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Lock "3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.222892] env[61906]: ERROR nova.compute.manager [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b01c48b8-7698-4df9-8653-500e06a0f61a, please check neutron logs for more information. [ 790.222892] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 790.222892] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 790.222892] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 790.222892] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 790.222892] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 790.222892] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 790.222892] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 790.222892] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 790.222892] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 790.222892] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 790.222892] env[61906]: ERROR nova.compute.manager raise self.value [ 790.222892] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 790.222892] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 790.222892] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 790.222892] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 790.223519] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 790.223519] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 790.223519] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b01c48b8-7698-4df9-8653-500e06a0f61a, please check neutron logs for more information. [ 790.223519] env[61906]: ERROR nova.compute.manager [ 790.223519] env[61906]: Traceback (most recent call last): [ 790.223519] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 790.223519] env[61906]: listener.cb(fileno) [ 790.223519] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 790.223519] env[61906]: result = function(*args, **kwargs) [ 790.223519] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 790.223519] env[61906]: return func(*args, **kwargs) [ 790.223519] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 790.223519] env[61906]: raise e [ 790.223519] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 790.223519] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 790.223519] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 790.223519] env[61906]: created_port_ids = self._update_ports_for_instance( [ 790.223519] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 790.223519] env[61906]: with excutils.save_and_reraise_exception(): [ 790.223519] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 790.223519] env[61906]: self.force_reraise() [ 790.223519] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 790.223519] env[61906]: raise self.value [ 790.223519] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 790.223519] env[61906]: updated_port = self._update_port( [ 790.223519] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 790.223519] env[61906]: _ensure_no_port_binding_failure(port) [ 790.223519] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 790.223519] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 790.224526] env[61906]: nova.exception.PortBindingFailed: Binding failed for port b01c48b8-7698-4df9-8653-500e06a0f61a, please check neutron logs for more information. [ 790.224526] env[61906]: Removing descriptor: 15 [ 790.224526] env[61906]: ERROR nova.compute.manager [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b01c48b8-7698-4df9-8653-500e06a0f61a, please check neutron logs for more information. [ 790.224526] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Traceback (most recent call last): [ 790.224526] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 790.224526] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] yield resources [ 790.224526] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 790.224526] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] self.driver.spawn(context, instance, image_meta, [ 790.224526] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 790.224526] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 790.224526] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 790.224526] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] vm_ref = self.build_virtual_machine(instance, [ 790.224952] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 790.224952] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] vif_infos = vmwarevif.get_vif_info(self._session, [ 790.224952] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 790.224952] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] for vif in network_info: [ 790.224952] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 790.224952] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] return self._sync_wrapper(fn, *args, **kwargs) [ 790.224952] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 790.224952] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] self.wait() [ 790.224952] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 790.224952] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] self[:] = self._gt.wait() [ 790.224952] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 790.224952] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] return self._exit_event.wait() [ 790.224952] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 790.225432] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] result = hub.switch() [ 790.225432] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 790.225432] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] return self.greenlet.switch() [ 790.225432] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 790.225432] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] result = function(*args, **kwargs) [ 790.225432] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 790.225432] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] return func(*args, **kwargs) [ 790.225432] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 790.225432] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] raise e [ 790.225432] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 790.225432] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] nwinfo = self.network_api.allocate_for_instance( [ 790.225432] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 790.225432] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] created_port_ids = self._update_ports_for_instance( [ 790.225884] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 790.225884] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] with excutils.save_and_reraise_exception(): [ 790.225884] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 790.225884] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] self.force_reraise() [ 790.225884] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 790.225884] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] raise self.value [ 790.225884] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 790.225884] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] updated_port = self._update_port( [ 790.225884] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 790.225884] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] _ensure_no_port_binding_failure(port) [ 790.225884] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 790.225884] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] raise exception.PortBindingFailed(port_id=port['id']) [ 790.226332] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] nova.exception.PortBindingFailed: Binding failed for port b01c48b8-7698-4df9-8653-500e06a0f61a, please check neutron logs for more information. [ 790.226332] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] [ 790.226332] env[61906]: INFO nova.compute.manager [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Terminating instance [ 790.226332] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Acquiring lock "refresh_cache-7ad136a2-d850-4b68-af0f-73badf65426c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.236812] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 790.361132] env[61906]: DEBUG nova.network.neutron [req-4a46a711-41dd-4aaa-a314-76f1278d0bc7 req-b40ee905-5b06-476b-80a6-870a9be1fb20 service nova] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 790.464476] env[61906]: DEBUG nova.network.neutron [req-4a46a711-41dd-4aaa-a314-76f1278d0bc7 req-b40ee905-5b06-476b-80a6-870a9be1fb20 service nova] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.741145] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance df9c9d62-8dc5-452c-85ed-846dcbd76014 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 790.968443] env[61906]: DEBUG oslo_concurrency.lockutils [req-4a46a711-41dd-4aaa-a314-76f1278d0bc7 req-b40ee905-5b06-476b-80a6-870a9be1fb20 service nova] Releasing lock "refresh_cache-7ad136a2-d850-4b68-af0f-73badf65426c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.968883] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Acquired lock "refresh_cache-7ad136a2-d850-4b68-af0f-73badf65426c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.969081] env[61906]: DEBUG nova.network.neutron [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 791.244751] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 52a387d2-4090-4eba-9a77-98c37f152b75 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 791.492563] env[61906]: DEBUG nova.network.neutron [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 791.600248] env[61906]: DEBUG nova.network.neutron [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.747999] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 34f333be-ea19-499e-ab89-e9f7663ae596 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 791.857754] env[61906]: DEBUG nova.compute.manager [req-7ff2ea84-f49d-4f8f-9189-5393f0f3bc1a req-0c37a0e4-f61c-4e00-9653-6b892e857ec5 service nova] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Received event network-vif-deleted-b01c48b8-7698-4df9-8653-500e06a0f61a {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 792.104718] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Releasing lock "refresh_cache-7ad136a2-d850-4b68-af0f-73badf65426c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.105919] env[61906]: DEBUG nova.compute.manager [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 792.105919] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 792.105919] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8925f254-8ad5-4346-a3b1-6dbc463ed4bc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.115591] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ded16b34-5eb2-4a94-a72f-a30e254c0a94 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.136599] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7ad136a2-d850-4b68-af0f-73badf65426c could not be found. [ 792.136813] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 792.136993] env[61906]: INFO nova.compute.manager [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 792.137276] env[61906]: DEBUG oslo.service.loopingcall [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 792.137507] env[61906]: DEBUG nova.compute.manager [-] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 792.137612] env[61906]: DEBUG nova.network.neutron [-] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 792.151556] env[61906]: DEBUG nova.network.neutron [-] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.250779] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 7183e141-9989-4a90-b2fb-2ca0d8e6cda9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 792.654256] env[61906]: DEBUG nova.network.neutron [-] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.753831] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance e0afe101-f4a4-4dc8-a666-9a7793dbf8a6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 793.156934] env[61906]: INFO nova.compute.manager [-] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Took 1.02 seconds to deallocate network for instance. [ 793.158843] env[61906]: DEBUG nova.compute.claims [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 793.159031] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.259148] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance a651bd36-b527-4743-bc6e-5534c629e409 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 793.765568] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 976021e6-0e38-4295-bdd4-83f00a7b7be0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 794.270726] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 7e05fee8-842f-4e03-853a-58bac5a9ac0e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 794.421026] env[61906]: DEBUG oslo_concurrency.lockutils [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Acquiring lock "75b9738e-e9b6-435a-88bb-851982828d36" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.421026] env[61906]: DEBUG oslo_concurrency.lockutils [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Lock "75b9738e-e9b6-435a-88bb-851982828d36" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.773108] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 9f1253e6-4727-4dca-bad2-04f0c5424cdb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 795.276033] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 80a6a0b8-17c3-46e8-8861-6d1759a19111 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 795.781211] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance d60575a6-af13-4d81-95d1-081d0f3e2ce6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 796.285934] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 796.286212] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 796.286366] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 796.536810] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7d26394-1fd5-4ba9-8c33-141d03bee9ac {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.544590] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ede6ba8-c6de-466e-9141-b484496b2d7e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.573445] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f2659e6-706c-48f5-b52d-a0703259c878 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.580638] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7c3f10d-4479-4214-a5ed-3adce3f1f778 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.595252] env[61906]: DEBUG nova.compute.provider_tree [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 797.099046] env[61906]: DEBUG nova.scheduler.client.report [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 797.603062] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 797.603338] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.042s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.603615] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.015s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.346559] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da3f363d-4ec4-4e81-84b1-31068d83c43c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.354341] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19c9b20c-79be-4236-a363-f5b549916d5a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.383170] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4c4fbde-1fe8-4e18-b9c9-71655e6cdbeb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.392975] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 798.393808] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Cleaning up deleted instances {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 798.401604] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-002a39d3-c58f-4ac2-beb7-7434eb73e426 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.415241] env[61906]: DEBUG nova.compute.provider_tree [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 798.898690] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] There are 3 instances to clean {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 798.899015] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 410b0308-1dca-4d11-91ab-f6d89bd565e3] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 798.918559] env[61906]: DEBUG nova.scheduler.client.report [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 799.402448] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 318a3224-a156-4354-a73b-460168c0eb9c] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 799.423379] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.820s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.423988] env[61906]: ERROR nova.compute.manager [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 40b6fa5b-eee9-46db-b7d0-e38d0315d3e5, please check neutron logs for more information. [ 799.423988] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Traceback (most recent call last): [ 799.423988] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 799.423988] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] self.driver.spawn(context, instance, image_meta, [ 799.423988] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 799.423988] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 799.423988] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 799.423988] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] vm_ref = self.build_virtual_machine(instance, [ 799.423988] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 799.423988] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] vif_infos = vmwarevif.get_vif_info(self._session, [ 799.423988] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 799.424262] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] for vif in network_info: [ 799.424262] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 799.424262] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] return self._sync_wrapper(fn, *args, **kwargs) [ 799.424262] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 799.424262] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] self.wait() [ 799.424262] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 799.424262] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] self[:] = self._gt.wait() [ 799.424262] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 799.424262] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] return self._exit_event.wait() [ 799.424262] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 799.424262] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] current.throw(*self._exc) [ 799.424262] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 799.424262] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] result = function(*args, **kwargs) [ 799.424584] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 799.424584] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] return func(*args, **kwargs) [ 799.424584] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 799.424584] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] raise e [ 799.424584] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 799.424584] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] nwinfo = self.network_api.allocate_for_instance( [ 799.424584] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 799.424584] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] created_port_ids = self._update_ports_for_instance( [ 799.424584] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 799.424584] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] with excutils.save_and_reraise_exception(): [ 799.424584] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 799.424584] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] self.force_reraise() [ 799.424584] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 799.424896] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] raise self.value [ 799.424896] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 799.424896] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] updated_port = self._update_port( [ 799.424896] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 799.424896] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] _ensure_no_port_binding_failure(port) [ 799.424896] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 799.424896] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] raise exception.PortBindingFailed(port_id=port['id']) [ 799.424896] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] nova.exception.PortBindingFailed: Binding failed for port 40b6fa5b-eee9-46db-b7d0-e38d0315d3e5, please check neutron logs for more information. [ 799.424896] env[61906]: ERROR nova.compute.manager [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] [ 799.424896] env[61906]: DEBUG nova.compute.utils [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Binding failed for port 40b6fa5b-eee9-46db-b7d0-e38d0315d3e5, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 799.425854] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.269s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.428701] env[61906]: DEBUG nova.compute.manager [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Build of instance 3df44b70-ee72-40d7-9129-b5e94e121b0a was re-scheduled: Binding failed for port 40b6fa5b-eee9-46db-b7d0-e38d0315d3e5, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 799.429119] env[61906]: DEBUG nova.compute.manager [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 799.429344] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "refresh_cache-3df44b70-ee72-40d7-9129-b5e94e121b0a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.429488] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired lock "refresh_cache-3df44b70-ee72-40d7-9129-b5e94e121b0a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.429645] env[61906]: DEBUG nova.network.neutron [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 799.906625] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: eb4cac2e-85d8-46b6-a4d4-ff12ee70e435] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 799.951735] env[61906]: DEBUG nova.network.neutron [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 800.048128] env[61906]: DEBUG nova.network.neutron [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.187202] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684a22a0-3e41-4100-b769-5f89cd9399e4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.195588] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9782ceb-4544-4734-a8c0-0556c13541ed {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.224526] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfc4c40d-795a-4c12-8a17-46b5a094ee1d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.231306] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4faecc1-579f-464d-b0da-8da58340aa16 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.244800] env[61906]: DEBUG nova.compute.provider_tree [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 800.411622] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 800.411622] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Cleaning up deleted instances with incomplete migration {{(pid=61906) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 800.552179] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Releasing lock "refresh_cache-3df44b70-ee72-40d7-9129-b5e94e121b0a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.552179] env[61906]: DEBUG nova.compute.manager [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 800.552292] env[61906]: DEBUG nova.compute.manager [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 800.552486] env[61906]: DEBUG nova.network.neutron [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 800.567288] env[61906]: DEBUG nova.network.neutron [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 800.748404] env[61906]: DEBUG nova.scheduler.client.report [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 800.914135] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 801.070258] env[61906]: DEBUG nova.network.neutron [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.254059] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.828s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.254786] env[61906]: ERROR nova.compute.manager [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 955316c4-ca43-4931-a1f2-cf43d26ffa46, please check neutron logs for more information. [ 801.254786] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Traceback (most recent call last): [ 801.254786] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 801.254786] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] self.driver.spawn(context, instance, image_meta, [ 801.254786] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 801.254786] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 801.254786] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 801.254786] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] vm_ref = self.build_virtual_machine(instance, [ 801.254786] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 801.254786] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] vif_infos = vmwarevif.get_vif_info(self._session, [ 801.254786] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 801.255139] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] for vif in network_info: [ 801.255139] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 801.255139] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] return self._sync_wrapper(fn, *args, **kwargs) [ 801.255139] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 801.255139] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] self.wait() [ 801.255139] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 801.255139] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] self[:] = self._gt.wait() [ 801.255139] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 801.255139] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] return self._exit_event.wait() [ 801.255139] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 801.255139] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] result = hub.switch() [ 801.255139] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 801.255139] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] return self.greenlet.switch() [ 801.255431] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 801.255431] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] result = function(*args, **kwargs) [ 801.255431] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 801.255431] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] return func(*args, **kwargs) [ 801.255431] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 801.255431] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] raise e [ 801.255431] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 801.255431] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] nwinfo = self.network_api.allocate_for_instance( [ 801.255431] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 801.255431] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] created_port_ids = self._update_ports_for_instance( [ 801.255431] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 801.255431] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] with excutils.save_and_reraise_exception(): [ 801.255431] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 801.255684] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] self.force_reraise() [ 801.255684] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 801.255684] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] raise self.value [ 801.255684] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 801.255684] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] updated_port = self._update_port( [ 801.255684] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 801.255684] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] _ensure_no_port_binding_failure(port) [ 801.255684] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 801.255684] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] raise exception.PortBindingFailed(port_id=port['id']) [ 801.255684] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] nova.exception.PortBindingFailed: Binding failed for port 955316c4-ca43-4931-a1f2-cf43d26ffa46, please check neutron logs for more information. [ 801.255684] env[61906]: ERROR nova.compute.manager [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] [ 801.255908] env[61906]: DEBUG nova.compute.utils [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Binding failed for port 955316c4-ca43-4931-a1f2-cf43d26ffa46, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 801.256680] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.077s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.258096] env[61906]: INFO nova.compute.claims [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 801.260770] env[61906]: DEBUG nova.compute.manager [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Build of instance a048328a-d4a8-420e-9cc8-047254d7d1bf was re-scheduled: Binding failed for port 955316c4-ca43-4931-a1f2-cf43d26ffa46, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 801.261197] env[61906]: DEBUG nova.compute.manager [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 801.261488] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Acquiring lock "refresh_cache-a048328a-d4a8-420e-9cc8-047254d7d1bf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.261620] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Acquired lock "refresh_cache-a048328a-d4a8-420e-9cc8-047254d7d1bf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.261776] env[61906]: DEBUG nova.network.neutron [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 801.573257] env[61906]: INFO nova.compute.manager [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 3df44b70-ee72-40d7-9129-b5e94e121b0a] Took 1.02 seconds to deallocate network for instance. [ 801.781074] env[61906]: DEBUG nova.network.neutron [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 801.856819] env[61906]: DEBUG nova.network.neutron [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.359821] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Releasing lock "refresh_cache-a048328a-d4a8-420e-9cc8-047254d7d1bf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.360084] env[61906]: DEBUG nova.compute.manager [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 802.360278] env[61906]: DEBUG nova.compute.manager [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 802.360438] env[61906]: DEBUG nova.network.neutron [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 802.387053] env[61906]: DEBUG nova.network.neutron [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 802.404900] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 802.404900] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 802.404900] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 802.404900] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 802.405311] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 802.405311] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 802.405437] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61906) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 802.405496] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 802.535028] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aa6218b-f725-4b55-b29d-e37cd9357af0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.543020] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f086c6e-c4d7-43af-815a-7d15f15835ea {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.573823] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42064342-03e6-4975-808a-5b38d93339cc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.580989] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8228024-819e-4140-a50c-6d5523072c30 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.596279] env[61906]: DEBUG nova.compute.provider_tree [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 802.614874] env[61906]: INFO nova.scheduler.client.report [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Deleted allocations for instance 3df44b70-ee72-40d7-9129-b5e94e121b0a [ 802.889786] env[61906]: DEBUG nova.network.neutron [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.908956] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.099680] env[61906]: DEBUG nova.scheduler.client.report [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 803.123414] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5332415-37ed-4132-9092-3b44ec38919b tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "3df44b70-ee72-40d7-9129-b5e94e121b0a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 195.969s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.392949] env[61906]: INFO nova.compute.manager [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] [instance: a048328a-d4a8-420e-9cc8-047254d7d1bf] Took 1.03 seconds to deallocate network for instance. [ 803.604687] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.348s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.605284] env[61906]: DEBUG nova.compute.manager [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 803.608509] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.545s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.628487] env[61906]: DEBUG nova.compute.manager [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 804.111138] env[61906]: DEBUG nova.compute.utils [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 804.112736] env[61906]: DEBUG nova.compute.manager [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Not allocating networking since 'none' was specified. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 804.149395] env[61906]: DEBUG oslo_concurrency.lockutils [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.362352] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea21f4ac-844f-4592-b6f7-664305b3831b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.370268] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99aefc6c-567a-4325-b5d5-46d5ecea6dcd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.400190] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95ef9c91-a31c-4480-b2a9-15a296c321f4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.410105] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d45c4a9b-6e61-48f5-8532-b234417041ee {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.423472] env[61906]: DEBUG nova.compute.provider_tree [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.426119] env[61906]: INFO nova.scheduler.client.report [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Deleted allocations for instance a048328a-d4a8-420e-9cc8-047254d7d1bf [ 804.617452] env[61906]: DEBUG nova.compute.manager [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 804.663914] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "cb34a0ce-d274-4167-8367-e19809cc9e4d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.663914] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "cb34a0ce-d274-4167-8367-e19809cc9e4d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.926888] env[61906]: DEBUG nova.scheduler.client.report [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 804.937292] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a37ad0c-0c6d-4356-a979-e3b20d0d90f8 tempest-FloatingIPsAssociationTestJSON-1734136560 tempest-FloatingIPsAssociationTestJSON-1734136560-project-member] Lock "a048328a-d4a8-420e-9cc8-047254d7d1bf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 189.170s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.431657] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.823s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.432366] env[61906]: ERROR nova.compute.manager [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 10d9cc70-a781-45a3-ba03-c4d9ed3c6d27, please check neutron logs for more information. [ 805.432366] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Traceback (most recent call last): [ 805.432366] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 805.432366] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] self.driver.spawn(context, instance, image_meta, [ 805.432366] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 805.432366] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 805.432366] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 805.432366] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] vm_ref = self.build_virtual_machine(instance, [ 805.432366] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 805.432366] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] vif_infos = vmwarevif.get_vif_info(self._session, [ 805.432366] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 805.432713] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] for vif in network_info: [ 805.432713] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 805.432713] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] return self._sync_wrapper(fn, *args, **kwargs) [ 805.432713] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 805.432713] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] self.wait() [ 805.432713] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 805.432713] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] self[:] = self._gt.wait() [ 805.432713] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 805.432713] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] return self._exit_event.wait() [ 805.432713] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 805.432713] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] current.throw(*self._exc) [ 805.432713] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 805.432713] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] result = function(*args, **kwargs) [ 805.433057] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 805.433057] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] return func(*args, **kwargs) [ 805.433057] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 805.433057] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] raise e [ 805.433057] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 805.433057] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] nwinfo = self.network_api.allocate_for_instance( [ 805.433057] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 805.433057] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] created_port_ids = self._update_ports_for_instance( [ 805.433057] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 805.433057] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] with excutils.save_and_reraise_exception(): [ 805.433057] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 805.433057] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] self.force_reraise() [ 805.433057] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 805.433382] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] raise self.value [ 805.433382] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 805.433382] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] updated_port = self._update_port( [ 805.433382] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 805.433382] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] _ensure_no_port_binding_failure(port) [ 805.433382] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 805.433382] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] raise exception.PortBindingFailed(port_id=port['id']) [ 805.433382] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] nova.exception.PortBindingFailed: Binding failed for port 10d9cc70-a781-45a3-ba03-c4d9ed3c6d27, please check neutron logs for more information. [ 805.433382] env[61906]: ERROR nova.compute.manager [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] [ 805.433382] env[61906]: DEBUG nova.compute.utils [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Binding failed for port 10d9cc70-a781-45a3-ba03-c4d9ed3c6d27, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 805.434408] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.749s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.438489] env[61906]: DEBUG nova.compute.manager [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Build of instance 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8 was re-scheduled: Binding failed for port 10d9cc70-a781-45a3-ba03-c4d9ed3c6d27, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 805.438489] env[61906]: DEBUG nova.compute.manager [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 805.438615] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Acquiring lock "refresh_cache-1b289a3d-7958-492b-ab9d-4ba8d3ffdae8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.438672] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Acquired lock "refresh_cache-1b289a3d-7958-492b-ab9d-4ba8d3ffdae8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.438827] env[61906]: DEBUG nova.network.neutron [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 805.440619] env[61906]: DEBUG nova.compute.manager [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 805.630634] env[61906]: DEBUG nova.compute.manager [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 805.656719] env[61906]: DEBUG nova.virt.hardware [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 805.657032] env[61906]: DEBUG nova.virt.hardware [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 805.657191] env[61906]: DEBUG nova.virt.hardware [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 805.657375] env[61906]: DEBUG nova.virt.hardware [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 805.657513] env[61906]: DEBUG nova.virt.hardware [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 805.657661] env[61906]: DEBUG nova.virt.hardware [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 805.658107] env[61906]: DEBUG nova.virt.hardware [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 805.658346] env[61906]: DEBUG nova.virt.hardware [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 805.658534] env[61906]: DEBUG nova.virt.hardware [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 805.658703] env[61906]: DEBUG nova.virt.hardware [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 805.658877] env[61906]: DEBUG nova.virt.hardware [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 805.659763] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-311d4aeb-a4f3-426d-a32d-62b2d4647893 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.668453] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1359801d-02cd-4ce5-b574-31baffb39b61 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.681687] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Instance VIF info [] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 805.687195] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Creating folder: Project (578d2968de15404fb57f5aa84dffb461). Parent ref: group-v284713. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 805.687195] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b2c5b9a4-1a94-4e80-8e50-8431992e2bdc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.696840] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Created folder: Project (578d2968de15404fb57f5aa84dffb461) in parent group-v284713. [ 805.697026] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Creating folder: Instances. Parent ref: group-v284730. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 805.697249] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bc44d096-9af9-4bfb-933d-af638d6695d6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.706498] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Created folder: Instances in parent group-v284730. [ 805.706498] env[61906]: DEBUG oslo.service.loopingcall [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 805.706619] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 805.706732] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-94ddf556-3249-4a37-a273-e829b4da33c7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.722676] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 805.722676] env[61906]: value = "task-1333139" [ 805.722676] env[61906]: _type = "Task" [ 805.722676] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.731234] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333139, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.968077] env[61906]: DEBUG nova.network.neutron [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 805.970563] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.059526] env[61906]: DEBUG nova.network.neutron [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.234280] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333139, 'name': CreateVM_Task, 'duration_secs': 0.257033} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.234447] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 806.234868] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.235034] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.235357] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 806.235614] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60c03758-a7b8-4a7f-84ba-9b6645fd281e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.238731] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f684c6b6-4445-4bd8-8355-f403c50b9f1b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.243757] env[61906]: DEBUG oslo_vmware.api [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Waiting for the task: (returnval){ [ 806.243757] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5257e276-48d8-ae32-74ae-0fabc737f4cf" [ 806.243757] env[61906]: _type = "Task" [ 806.243757] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.249856] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-901b295c-ec06-4e6f-a9a6-e7ce420e9008 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.259111] env[61906]: DEBUG oslo_vmware.api [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5257e276-48d8-ae32-74ae-0fabc737f4cf, 'name': SearchDatastore_Task, 'duration_secs': 0.010883} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.281995] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.282249] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 806.282496] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.282638] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.282821] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 806.283249] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0ff85abe-1a1f-4ddd-8890-f4262c4f8db2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.285481] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d9cb110-4333-48eb-8057-1e282197b386 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.292885] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06b4fd6a-036a-4706-af4d-19ff98d273f9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.299872] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 806.300193] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 806.302123] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f8b37d2-dbf0-4440-b4bd-1083abd79805 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.311768] env[61906]: DEBUG nova.compute.provider_tree [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.316261] env[61906]: DEBUG oslo_vmware.api [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Waiting for the task: (returnval){ [ 806.316261] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]529b0610-4435-d6b6-012f-11602d6a09bd" [ 806.316261] env[61906]: _type = "Task" [ 806.316261] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.324995] env[61906]: DEBUG oslo_vmware.api [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]529b0610-4435-d6b6-012f-11602d6a09bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.564125] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Releasing lock "refresh_cache-1b289a3d-7958-492b-ab9d-4ba8d3ffdae8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.564125] env[61906]: DEBUG nova.compute.manager [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 806.564125] env[61906]: DEBUG nova.compute.manager [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 806.564125] env[61906]: DEBUG nova.network.neutron [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 806.584551] env[61906]: DEBUG nova.network.neutron [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 806.816528] env[61906]: DEBUG nova.scheduler.client.report [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 806.830620] env[61906]: DEBUG oslo_vmware.api [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]529b0610-4435-d6b6-012f-11602d6a09bd, 'name': SearchDatastore_Task, 'duration_secs': 0.009063} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.831476] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-396dffe6-592d-4a7f-b22e-d9c81b4f29b1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.838291] env[61906]: DEBUG oslo_vmware.api [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Waiting for the task: (returnval){ [ 806.838291] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52b66db4-6dee-326f-89a9-9ca776a4b883" [ 806.838291] env[61906]: _type = "Task" [ 806.838291] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.847320] env[61906]: DEBUG oslo_vmware.api [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52b66db4-6dee-326f-89a9-9ca776a4b883, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.087088] env[61906]: DEBUG nova.network.neutron [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.321968] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.886s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.321968] env[61906]: ERROR nova.compute.manager [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 558793cc-c0df-4085-9518-c5b51652e2c5, please check neutron logs for more information. [ 807.321968] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Traceback (most recent call last): [ 807.321968] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 807.321968] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] self.driver.spawn(context, instance, image_meta, [ 807.321968] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 807.321968] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] self._vmops.spawn(context, instance, image_meta, injected_files, [ 807.321968] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 807.321968] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] vm_ref = self.build_virtual_machine(instance, [ 807.322314] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 807.322314] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] vif_infos = vmwarevif.get_vif_info(self._session, [ 807.322314] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 807.322314] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] for vif in network_info: [ 807.322314] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 807.322314] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] return self._sync_wrapper(fn, *args, **kwargs) [ 807.322314] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 807.322314] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] self.wait() [ 807.322314] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 807.322314] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] self[:] = self._gt.wait() [ 807.322314] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 807.322314] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] return self._exit_event.wait() [ 807.322314] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 807.322752] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] result = hub.switch() [ 807.322752] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 807.322752] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] return self.greenlet.switch() [ 807.322752] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 807.322752] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] result = function(*args, **kwargs) [ 807.322752] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 807.322752] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] return func(*args, **kwargs) [ 807.322752] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 807.322752] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] raise e [ 807.322752] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 807.322752] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] nwinfo = self.network_api.allocate_for_instance( [ 807.322752] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 807.322752] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] created_port_ids = self._update_ports_for_instance( [ 807.323151] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 807.323151] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] with excutils.save_and_reraise_exception(): [ 807.323151] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 807.323151] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] self.force_reraise() [ 807.323151] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 807.323151] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] raise self.value [ 807.323151] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 807.323151] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] updated_port = self._update_port( [ 807.323151] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 807.323151] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] _ensure_no_port_binding_failure(port) [ 807.323151] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 807.323151] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] raise exception.PortBindingFailed(port_id=port['id']) [ 807.323468] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] nova.exception.PortBindingFailed: Binding failed for port 558793cc-c0df-4085-9518-c5b51652e2c5, please check neutron logs for more information. [ 807.323468] env[61906]: ERROR nova.compute.manager [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] [ 807.323468] env[61906]: DEBUG nova.compute.utils [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Binding failed for port 558793cc-c0df-4085-9518-c5b51652e2c5, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 807.323894] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.471s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.327996] env[61906]: DEBUG nova.compute.manager [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Build of instance 830e7d50-5870-4e1f-8f2f-26840ba06d33 was re-scheduled: Binding failed for port 558793cc-c0df-4085-9518-c5b51652e2c5, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 807.328854] env[61906]: DEBUG nova.compute.manager [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 807.329108] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Acquiring lock "refresh_cache-830e7d50-5870-4e1f-8f2f-26840ba06d33" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.329264] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Acquired lock "refresh_cache-830e7d50-5870-4e1f-8f2f-26840ba06d33" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.329577] env[61906]: DEBUG nova.network.neutron [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 807.350522] env[61906]: DEBUG oslo_vmware.api [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52b66db4-6dee-326f-89a9-9ca776a4b883, 'name': SearchDatastore_Task, 'duration_secs': 0.008859} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.350887] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.351189] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 7e452659-0a5e-4740-b7ed-1fd7b1917dac/7e452659-0a5e-4740-b7ed-1fd7b1917dac.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 807.351564] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ae2ba3f8-8a33-4720-895f-056cd3f4e085 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.359121] env[61906]: DEBUG oslo_vmware.api [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Waiting for the task: (returnval){ [ 807.359121] env[61906]: value = "task-1333140" [ 807.359121] env[61906]: _type = "Task" [ 807.359121] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.366858] env[61906]: DEBUG oslo_vmware.api [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333140, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.590110] env[61906]: INFO nova.compute.manager [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] [instance: 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8] Took 1.03 seconds to deallocate network for instance. [ 807.849931] env[61906]: DEBUG nova.network.neutron [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 807.867860] env[61906]: DEBUG oslo_vmware.api [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333140, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.450871} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.869981] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 7e452659-0a5e-4740-b7ed-1fd7b1917dac/7e452659-0a5e-4740-b7ed-1fd7b1917dac.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 807.870186] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 807.870454] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-abf40375-c430-415f-a998-cf3f05c04e6f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.878466] env[61906]: DEBUG oslo_vmware.api [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Waiting for the task: (returnval){ [ 807.878466] env[61906]: value = "task-1333141" [ 807.878466] env[61906]: _type = "Task" [ 807.878466] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.888431] env[61906]: DEBUG oslo_vmware.api [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333141, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.907439] env[61906]: DEBUG nova.network.neutron [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.238322] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-430c5679-69bf-48de-baeb-a86e11594ce0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.246814] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cdc1706-5b8f-4925-84c1-787840d54266 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.278609] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a4869ad-1f93-4f09-af56-ea80b0ff5474 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.286151] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1850dfc-9cd7-40d2-8ac0-1a358b95ca75 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.300161] env[61906]: DEBUG nova.compute.provider_tree [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.388780] env[61906]: DEBUG oslo_vmware.api [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333141, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064144} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.389246] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 808.390077] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8c6bf92-dffd-4175-bd6b-5d6b05f9521d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.409438] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] 7e452659-0a5e-4740-b7ed-1fd7b1917dac/7e452659-0a5e-4740-b7ed-1fd7b1917dac.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 808.409921] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Releasing lock "refresh_cache-830e7d50-5870-4e1f-8f2f-26840ba06d33" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.410143] env[61906]: DEBUG nova.compute.manager [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 808.410310] env[61906]: DEBUG nova.compute.manager [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 808.410474] env[61906]: DEBUG nova.network.neutron [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 808.412025] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed5b8563-05e3-4c55-95ab-7bad1d87e13b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.431626] env[61906]: DEBUG oslo_vmware.api [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Waiting for the task: (returnval){ [ 808.431626] env[61906]: value = "task-1333142" [ 808.431626] env[61906]: _type = "Task" [ 808.431626] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.439501] env[61906]: DEBUG oslo_vmware.api [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333142, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.440286] env[61906]: DEBUG nova.network.neutron [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.624754] env[61906]: INFO nova.scheduler.client.report [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Deleted allocations for instance 1b289a3d-7958-492b-ab9d-4ba8d3ffdae8 [ 808.803207] env[61906]: DEBUG nova.scheduler.client.report [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 808.942809] env[61906]: DEBUG nova.network.neutron [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.944235] env[61906]: DEBUG oslo_vmware.api [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333142, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.136520] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86b8a198-fd6e-4d81-a076-d036e29e6bd2 tempest-ServerGroupTestJSON-392548268 tempest-ServerGroupTestJSON-392548268-project-member] Lock "1b289a3d-7958-492b-ab9d-4ba8d3ffdae8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 172.271s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.309161] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.984s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.309161] env[61906]: ERROR nova.compute.manager [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5d3b6ed6-477e-4457-94a1-2b812721d3e3, please check neutron logs for more information. [ 809.309161] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Traceback (most recent call last): [ 809.309161] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 809.309161] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] self.driver.spawn(context, instance, image_meta, [ 809.309161] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 809.309161] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 809.309161] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 809.309161] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] vm_ref = self.build_virtual_machine(instance, [ 809.309392] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 809.309392] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] vif_infos = vmwarevif.get_vif_info(self._session, [ 809.309392] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 809.309392] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] for vif in network_info: [ 809.309392] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 809.309392] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] return self._sync_wrapper(fn, *args, **kwargs) [ 809.309392] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 809.309392] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] self.wait() [ 809.309392] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 809.309392] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] self[:] = self._gt.wait() [ 809.309392] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 809.309392] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] return self._exit_event.wait() [ 809.309392] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 809.309679] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] current.throw(*self._exc) [ 809.309679] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 809.309679] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] result = function(*args, **kwargs) [ 809.309679] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 809.309679] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] return func(*args, **kwargs) [ 809.309679] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 809.309679] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] raise e [ 809.309679] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 809.309679] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] nwinfo = self.network_api.allocate_for_instance( [ 809.309679] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 809.309679] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] created_port_ids = self._update_ports_for_instance( [ 809.309679] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 809.309679] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] with excutils.save_and_reraise_exception(): [ 809.309951] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 809.309951] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] self.force_reraise() [ 809.309951] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 809.309951] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] raise self.value [ 809.309951] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 809.309951] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] updated_port = self._update_port( [ 809.309951] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 809.309951] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] _ensure_no_port_binding_failure(port) [ 809.309951] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 809.309951] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] raise exception.PortBindingFailed(port_id=port['id']) [ 809.309951] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] nova.exception.PortBindingFailed: Binding failed for port 5d3b6ed6-477e-4457-94a1-2b812721d3e3, please check neutron logs for more information. [ 809.309951] env[61906]: ERROR nova.compute.manager [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] [ 809.313295] env[61906]: DEBUG nova.compute.utils [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Binding failed for port 5d3b6ed6-477e-4457-94a1-2b812721d3e3, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 809.314709] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.077s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.316695] env[61906]: INFO nova.compute.claims [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 809.323936] env[61906]: DEBUG nova.compute.manager [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Build of instance 5ab321ec-655b-4f68-91aa-9b3db8c692f4 was re-scheduled: Binding failed for port 5d3b6ed6-477e-4457-94a1-2b812721d3e3, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 809.324484] env[61906]: DEBUG nova.compute.manager [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 809.324764] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Acquiring lock "refresh_cache-5ab321ec-655b-4f68-91aa-9b3db8c692f4" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.324916] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Acquired lock "refresh_cache-5ab321ec-655b-4f68-91aa-9b3db8c692f4" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.325107] env[61906]: DEBUG nova.network.neutron [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 809.442990] env[61906]: DEBUG oslo_vmware.api [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333142, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.445791] env[61906]: INFO nova.compute.manager [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] [instance: 830e7d50-5870-4e1f-8f2f-26840ba06d33] Took 1.04 seconds to deallocate network for instance. [ 809.642301] env[61906]: DEBUG nova.compute.manager [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 809.859554] env[61906]: DEBUG nova.network.neutron [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 809.942723] env[61906]: DEBUG oslo_vmware.api [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333142, 'name': ReconfigVM_Task, 'duration_secs': 1.286612} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.943009] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Reconfigured VM instance instance-00000033 to attach disk [datastore2] 7e452659-0a5e-4740-b7ed-1fd7b1917dac/7e452659-0a5e-4740-b7ed-1fd7b1917dac.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 809.943594] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c91838aa-0a7e-4b38-a6f2-9a9d60d2beef {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.949356] env[61906]: DEBUG oslo_vmware.api [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Waiting for the task: (returnval){ [ 809.949356] env[61906]: value = "task-1333143" [ 809.949356] env[61906]: _type = "Task" [ 809.949356] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.963900] env[61906]: DEBUG oslo_vmware.api [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333143, 'name': Rename_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.976667] env[61906]: DEBUG nova.network.neutron [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.166258] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.349272] env[61906]: DEBUG nova.scheduler.client.report [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Refreshing inventories for resource provider a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 810.365134] env[61906]: DEBUG nova.scheduler.client.report [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Updating ProviderTree inventory for provider a5b57df7-a16e-44f9-9b2d-23c518860263 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 810.365134] env[61906]: DEBUG nova.compute.provider_tree [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 810.377997] env[61906]: DEBUG nova.scheduler.client.report [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Refreshing aggregate associations for resource provider a5b57df7-a16e-44f9-9b2d-23c518860263, aggregates: None {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 810.397196] env[61906]: DEBUG nova.scheduler.client.report [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Refreshing trait associations for resource provider a5b57df7-a16e-44f9-9b2d-23c518860263, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64 {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 810.469151] env[61906]: DEBUG oslo_vmware.api [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333143, 'name': Rename_Task, 'duration_secs': 0.19832} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.469468] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 810.469790] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-af7173b1-db6b-4f1f-8c37-6b9aa447e849 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.479046] env[61906]: DEBUG oslo_vmware.api [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Waiting for the task: (returnval){ [ 810.479046] env[61906]: value = "task-1333144" [ 810.479046] env[61906]: _type = "Task" [ 810.479046] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.481464] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Releasing lock "refresh_cache-5ab321ec-655b-4f68-91aa-9b3db8c692f4" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.481586] env[61906]: DEBUG nova.compute.manager [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 810.481768] env[61906]: DEBUG nova.compute.manager [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 810.481934] env[61906]: DEBUG nova.network.neutron [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 810.493161] env[61906]: DEBUG oslo_vmware.api [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333144, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.494396] env[61906]: INFO nova.scheduler.client.report [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Deleted allocations for instance 830e7d50-5870-4e1f-8f2f-26840ba06d33 [ 810.538876] env[61906]: DEBUG nova.network.neutron [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 810.780892] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0ccc09e-8ebb-4b95-8e61-ca995eac18d1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.789115] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fab3d3e-6731-4903-b8d4-47cdde76c2d4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.828477] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b92e20-fc1a-4f48-9d75-66315a2fe7f2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.839184] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58fd867d-4ed4-463b-a996-81a6e39e6e01 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.859192] env[61906]: DEBUG nova.compute.provider_tree [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.996176] env[61906]: DEBUG oslo_vmware.api [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333144, 'name': PowerOnVM_Task, 'duration_secs': 0.420634} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.996505] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 810.996711] env[61906]: INFO nova.compute.manager [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Took 5.37 seconds to spawn the instance on the hypervisor. [ 810.996888] env[61906]: DEBUG nova.compute.manager [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 810.997702] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92bedc90-857e-4496-abae-c7add041ce5c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.003927] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f9a19fcb-589e-4c00-ab3b-effaaaebf1cf tempest-ServerDiagnosticsNegativeTest-182162039 tempest-ServerDiagnosticsNegativeTest-182162039-project-member] Lock "830e7d50-5870-4e1f-8f2f-26840ba06d33" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 169.165s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.041981] env[61906]: DEBUG nova.network.neutron [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.362635] env[61906]: DEBUG nova.scheduler.client.report [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 811.518926] env[61906]: DEBUG nova.compute.manager [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 811.529058] env[61906]: INFO nova.compute.manager [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Took 34.37 seconds to build instance. [ 811.545041] env[61906]: INFO nova.compute.manager [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] [instance: 5ab321ec-655b-4f68-91aa-9b3db8c692f4] Took 1.06 seconds to deallocate network for instance. [ 811.868430] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.554s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.868969] env[61906]: DEBUG nova.compute.manager [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 811.874428] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.775s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.877207] env[61906]: INFO nova.compute.claims [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 812.032274] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d8dc2689-ab87-4414-8aa0-3e5ce6ddfdc1 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Lock "7e452659-0a5e-4740-b7ed-1fd7b1917dac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 161.783s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.057758] env[61906]: DEBUG oslo_concurrency.lockutils [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.387162] env[61906]: DEBUG nova.compute.utils [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 812.389608] env[61906]: DEBUG nova.compute.manager [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 812.390374] env[61906]: DEBUG nova.network.neutron [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 812.396518] env[61906]: INFO nova.compute.manager [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Rebuilding instance [ 812.444956] env[61906]: DEBUG nova.policy [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd660b01922b6407cb85df643be9a2366', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6395ad275c56495fadd6593bd3429eed', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 812.446772] env[61906]: DEBUG nova.compute.manager [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 812.448218] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42e2e084-653a-4852-9f29-f7b9e9bb9e78 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.537029] env[61906]: DEBUG nova.compute.manager [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 812.605393] env[61906]: INFO nova.scheduler.client.report [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Deleted allocations for instance 5ab321ec-655b-4f68-91aa-9b3db8c692f4 [ 812.897604] env[61906]: DEBUG nova.network.neutron [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Successfully created port: 04cfd791-c3c1-4e94-8768-90499c8e9ca0 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 812.900118] env[61906]: DEBUG nova.compute.manager [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 812.961805] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 812.962818] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bfcc056a-6530-425c-bd2f-535e8c3a25c3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.969918] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Waiting for the task: (returnval){ [ 812.969918] env[61906]: value = "task-1333148" [ 812.969918] env[61906]: _type = "Task" [ 812.969918] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.979016] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333148, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.061517] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.116509] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e9ff132-a335-4e3d-9d03-c211240f69d0 tempest-MigrationsAdminTest-1370676927 tempest-MigrationsAdminTest-1370676927-project-member] Lock "5ab321ec-655b-4f68-91aa-9b3db8c692f4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 169.232s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.226350] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52067d93-63f8-4b02-bed2-0534595052c8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.234281] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1796e59-d97d-4d05-91ca-12508e1b5d0c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.268043] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d8a82b-d084-4c40-a229-06a7b84ee179 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.275483] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d12f6fc2-986b-4dc7-801b-44f004eea57a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.289708] env[61906]: DEBUG nova.compute.provider_tree [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.483349] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333148, 'name': PowerOffVM_Task, 'duration_secs': 0.217019} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.483759] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 813.484089] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 813.486588] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbfdbfdd-4c55-47fd-9f44-81ce9e9a91e8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.494396] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 813.494670] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d4d70882-2e7a-4d08-8b11-fa1a5ce8418d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.523299] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 813.523299] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 813.523299] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Deleting the datastore file [datastore2] 7e452659-0a5e-4740-b7ed-1fd7b1917dac {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 813.523299] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-61a4f95a-367a-4ae6-9bab-a9d5475d0f5e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.529651] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Waiting for the task: (returnval){ [ 813.529651] env[61906]: value = "task-1333150" [ 813.529651] env[61906]: _type = "Task" [ 813.529651] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.543328] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333150, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.619395] env[61906]: DEBUG nova.compute.manager [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 813.746184] env[61906]: DEBUG nova.compute.manager [req-3cd98a83-5ae3-4de2-a1eb-f8ffaa9d5f5e req-b27bda5e-4971-4bc4-9e27-6506dab1d560 service nova] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Received event network-changed-04cfd791-c3c1-4e94-8768-90499c8e9ca0 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 813.746497] env[61906]: DEBUG nova.compute.manager [req-3cd98a83-5ae3-4de2-a1eb-f8ffaa9d5f5e req-b27bda5e-4971-4bc4-9e27-6506dab1d560 service nova] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Refreshing instance network info cache due to event network-changed-04cfd791-c3c1-4e94-8768-90499c8e9ca0. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 813.746623] env[61906]: DEBUG oslo_concurrency.lockutils [req-3cd98a83-5ae3-4de2-a1eb-f8ffaa9d5f5e req-b27bda5e-4971-4bc4-9e27-6506dab1d560 service nova] Acquiring lock "refresh_cache-b935c7ba-7716-443b-b5da-9ee03cef793a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.746766] env[61906]: DEBUG oslo_concurrency.lockutils [req-3cd98a83-5ae3-4de2-a1eb-f8ffaa9d5f5e req-b27bda5e-4971-4bc4-9e27-6506dab1d560 service nova] Acquired lock "refresh_cache-b935c7ba-7716-443b-b5da-9ee03cef793a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.746935] env[61906]: DEBUG nova.network.neutron [req-3cd98a83-5ae3-4de2-a1eb-f8ffaa9d5f5e req-b27bda5e-4971-4bc4-9e27-6506dab1d560 service nova] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Refreshing network info cache for port 04cfd791-c3c1-4e94-8768-90499c8e9ca0 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 813.794246] env[61906]: DEBUG nova.scheduler.client.report [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 813.917972] env[61906]: DEBUG nova.compute.manager [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 813.950670] env[61906]: DEBUG nova.virt.hardware [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 813.950670] env[61906]: DEBUG nova.virt.hardware [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 813.950670] env[61906]: DEBUG nova.virt.hardware [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 813.950898] env[61906]: DEBUG nova.virt.hardware [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 813.950898] env[61906]: DEBUG nova.virt.hardware [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 813.950898] env[61906]: DEBUG nova.virt.hardware [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 813.951124] env[61906]: DEBUG nova.virt.hardware [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 813.951421] env[61906]: DEBUG nova.virt.hardware [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 813.951731] env[61906]: DEBUG nova.virt.hardware [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 813.952245] env[61906]: DEBUG nova.virt.hardware [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 813.952528] env[61906]: DEBUG nova.virt.hardware [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 813.954321] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5094d96-2074-491d-88a5-3334898c963a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.963364] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9834eb6-d364-4def-88f7-f29f8248d9ee {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.773944] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.899s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.774450] env[61906]: DEBUG nova.compute.manager [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 814.778477] env[61906]: ERROR nova.compute.manager [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 04cfd791-c3c1-4e94-8768-90499c8e9ca0, please check neutron logs for more information. [ 814.778477] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 814.778477] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 814.778477] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 814.778477] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 814.778477] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 814.778477] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 814.778477] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 814.778477] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 814.778477] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 814.778477] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 814.778477] env[61906]: ERROR nova.compute.manager raise self.value [ 814.778477] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 814.778477] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 814.778477] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 814.778477] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 814.778950] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 814.778950] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 814.778950] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 04cfd791-c3c1-4e94-8768-90499c8e9ca0, please check neutron logs for more information. [ 814.778950] env[61906]: ERROR nova.compute.manager [ 814.778950] env[61906]: Traceback (most recent call last): [ 814.778950] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 814.778950] env[61906]: listener.cb(fileno) [ 814.778950] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 814.778950] env[61906]: result = function(*args, **kwargs) [ 814.778950] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 814.778950] env[61906]: return func(*args, **kwargs) [ 814.778950] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 814.778950] env[61906]: raise e [ 814.778950] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 814.778950] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 814.778950] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 814.778950] env[61906]: created_port_ids = self._update_ports_for_instance( [ 814.778950] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 814.778950] env[61906]: with excutils.save_and_reraise_exception(): [ 814.778950] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 814.778950] env[61906]: self.force_reraise() [ 814.778950] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 814.778950] env[61906]: raise self.value [ 814.778950] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 814.778950] env[61906]: updated_port = self._update_port( [ 814.778950] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 814.778950] env[61906]: _ensure_no_port_binding_failure(port) [ 814.778950] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 814.778950] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 814.779790] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 04cfd791-c3c1-4e94-8768-90499c8e9ca0, please check neutron logs for more information. [ 814.779790] env[61906]: Removing descriptor: 15 [ 814.782187] env[61906]: DEBUG oslo_concurrency.lockutils [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.850s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.785217] env[61906]: INFO nova.compute.claims [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 814.787061] env[61906]: ERROR nova.compute.manager [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 04cfd791-c3c1-4e94-8768-90499c8e9ca0, please check neutron logs for more information. [ 814.787061] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Traceback (most recent call last): [ 814.787061] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 814.787061] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] yield resources [ 814.787061] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 814.787061] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] self.driver.spawn(context, instance, image_meta, [ 814.787061] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 814.787061] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 814.787061] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 814.787061] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] vm_ref = self.build_virtual_machine(instance, [ 814.787061] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 814.787345] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] vif_infos = vmwarevif.get_vif_info(self._session, [ 814.787345] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 814.787345] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] for vif in network_info: [ 814.787345] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 814.787345] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] return self._sync_wrapper(fn, *args, **kwargs) [ 814.787345] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 814.787345] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] self.wait() [ 814.787345] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 814.787345] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] self[:] = self._gt.wait() [ 814.787345] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 814.787345] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] return self._exit_event.wait() [ 814.787345] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 814.787345] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] result = hub.switch() [ 814.787629] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 814.787629] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] return self.greenlet.switch() [ 814.787629] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 814.787629] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] result = function(*args, **kwargs) [ 814.787629] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 814.787629] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] return func(*args, **kwargs) [ 814.787629] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 814.787629] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] raise e [ 814.787629] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 814.787629] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] nwinfo = self.network_api.allocate_for_instance( [ 814.787629] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 814.787629] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] created_port_ids = self._update_ports_for_instance( [ 814.787629] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 814.787910] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] with excutils.save_and_reraise_exception(): [ 814.787910] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 814.787910] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] self.force_reraise() [ 814.787910] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 814.787910] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] raise self.value [ 814.787910] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 814.787910] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] updated_port = self._update_port( [ 814.787910] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 814.787910] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] _ensure_no_port_binding_failure(port) [ 814.787910] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 814.787910] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] raise exception.PortBindingFailed(port_id=port['id']) [ 814.787910] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] nova.exception.PortBindingFailed: Binding failed for port 04cfd791-c3c1-4e94-8768-90499c8e9ca0, please check neutron logs for more information. [ 814.787910] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] [ 814.788208] env[61906]: INFO nova.compute.manager [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Terminating instance [ 814.788935] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Acquiring lock "refresh_cache-b935c7ba-7716-443b-b5da-9ee03cef793a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.796408] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333150, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.090544} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.796641] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 814.796825] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 814.796997] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 814.800204] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.810107] env[61906]: DEBUG nova.network.neutron [req-3cd98a83-5ae3-4de2-a1eb-f8ffaa9d5f5e req-b27bda5e-4971-4bc4-9e27-6506dab1d560 service nova] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 814.914121] env[61906]: DEBUG nova.network.neutron [req-3cd98a83-5ae3-4de2-a1eb-f8ffaa9d5f5e req-b27bda5e-4971-4bc4-9e27-6506dab1d560 service nova] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.279293] env[61906]: DEBUG nova.compute.utils [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 815.280743] env[61906]: DEBUG nova.compute.manager [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 815.280907] env[61906]: DEBUG nova.network.neutron [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 815.333054] env[61906]: DEBUG nova.policy [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '43637caec41a4f30981c055ff2a36110', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '44eb493d234a4647b98e6fff00e739dc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 815.416303] env[61906]: DEBUG oslo_concurrency.lockutils [req-3cd98a83-5ae3-4de2-a1eb-f8ffaa9d5f5e req-b27bda5e-4971-4bc4-9e27-6506dab1d560 service nova] Releasing lock "refresh_cache-b935c7ba-7716-443b-b5da-9ee03cef793a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.416709] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Acquired lock "refresh_cache-b935c7ba-7716-443b-b5da-9ee03cef793a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.416898] env[61906]: DEBUG nova.network.neutron [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 815.714162] env[61906]: DEBUG nova.network.neutron [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Successfully created port: 88d10958-f4d5-4378-98b5-dae42820a117 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 815.786974] env[61906]: DEBUG nova.compute.manager [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 815.822867] env[61906]: DEBUG nova.compute.manager [req-998e4b70-2c2b-4253-888a-a40016ffb328 req-f0215ee2-a43a-41eb-9a4e-e7605893cfd4 service nova] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Received event network-vif-deleted-04cfd791-c3c1-4e94-8768-90499c8e9ca0 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 815.851792] env[61906]: DEBUG nova.virt.hardware [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 815.852078] env[61906]: DEBUG nova.virt.hardware [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 815.852259] env[61906]: DEBUG nova.virt.hardware [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 815.852454] env[61906]: DEBUG nova.virt.hardware [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 815.852650] env[61906]: DEBUG nova.virt.hardware [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 815.852785] env[61906]: DEBUG nova.virt.hardware [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 815.852992] env[61906]: DEBUG nova.virt.hardware [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 815.853406] env[61906]: DEBUG nova.virt.hardware [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 815.853595] env[61906]: DEBUG nova.virt.hardware [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 815.853760] env[61906]: DEBUG nova.virt.hardware [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 815.854057] env[61906]: DEBUG nova.virt.hardware [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 815.855494] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f4ea523-e408-446d-8363-d5467e0cc34c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.869251] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e779f1b-52b8-4ad2-8033-cc368fea5fb1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.884611] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Instance VIF info [] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 815.890377] env[61906]: DEBUG oslo.service.loopingcall [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 815.893236] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 815.893657] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2d4237a7-85fe-40e0-bece-a6257ac9aed2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.914159] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 815.914159] env[61906]: value = "task-1333152" [ 815.914159] env[61906]: _type = "Task" [ 815.914159] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.926541] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333152, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.944554] env[61906]: DEBUG nova.network.neutron [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 816.088408] env[61906]: DEBUG nova.network.neutron [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.143144] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18ad2efe-2103-4a71-84bf-02855b847fdd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.150457] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eddd779b-80e0-48a7-b9e6-0718495b26fc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.184130] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1cf5b7e-f855-47dd-9dff-a0d3c346b978 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.191967] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45504f27-5e42-4ec9-814a-96ac7dac913b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.205829] env[61906]: DEBUG nova.compute.provider_tree [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 816.423545] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333152, 'name': CreateVM_Task, 'duration_secs': 0.327023} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.423718] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 816.424264] env[61906]: DEBUG oslo_concurrency.lockutils [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.424309] env[61906]: DEBUG oslo_concurrency.lockutils [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.424607] env[61906]: DEBUG oslo_concurrency.lockutils [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 816.424847] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dcd3e5e1-54e6-48ae-a712-2faf630f726c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.429032] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Waiting for the task: (returnval){ [ 816.429032] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52946a5e-a09c-7b0c-3c88-4398e205d69b" [ 816.429032] env[61906]: _type = "Task" [ 816.429032] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.437719] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52946a5e-a09c-7b0c-3c88-4398e205d69b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.592366] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Releasing lock "refresh_cache-b935c7ba-7716-443b-b5da-9ee03cef793a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.593049] env[61906]: DEBUG nova.compute.manager [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 816.593441] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 816.593766] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-24a16205-8341-4497-bb00-943b6993b730 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.604732] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bab8ce1-15d6-4c6e-a03e-7915dfeb24ec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.638468] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b935c7ba-7716-443b-b5da-9ee03cef793a could not be found. [ 816.638468] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 816.638468] env[61906]: INFO nova.compute.manager [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Took 0.05 seconds to destroy the instance on the hypervisor. [ 816.638602] env[61906]: DEBUG oslo.service.loopingcall [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.638781] env[61906]: DEBUG nova.compute.manager [-] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 816.638877] env[61906]: DEBUG nova.network.neutron [-] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 816.663507] env[61906]: DEBUG nova.network.neutron [-] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 816.709393] env[61906]: DEBUG nova.scheduler.client.report [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 816.798280] env[61906]: DEBUG nova.compute.manager [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 816.833585] env[61906]: DEBUG nova.virt.hardware [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 816.834063] env[61906]: DEBUG nova.virt.hardware [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 816.835306] env[61906]: DEBUG nova.virt.hardware [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 816.836060] env[61906]: DEBUG nova.virt.hardware [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 816.836283] env[61906]: DEBUG nova.virt.hardware [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 816.836449] env[61906]: DEBUG nova.virt.hardware [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 816.836660] env[61906]: DEBUG nova.virt.hardware [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 816.836831] env[61906]: DEBUG nova.virt.hardware [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 816.838038] env[61906]: DEBUG nova.virt.hardware [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 816.838234] env[61906]: DEBUG nova.virt.hardware [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 816.838452] env[61906]: DEBUG nova.virt.hardware [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 816.839352] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9acead1f-6f2b-4196-8a0b-fefed5d38c93 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.842814] env[61906]: ERROR nova.compute.manager [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 88d10958-f4d5-4378-98b5-dae42820a117, please check neutron logs for more information. [ 816.842814] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 816.842814] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 816.842814] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 816.842814] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 816.842814] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 816.842814] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 816.842814] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 816.842814] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 816.842814] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 816.842814] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 816.842814] env[61906]: ERROR nova.compute.manager raise self.value [ 816.842814] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 816.842814] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 816.842814] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 816.842814] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 816.843310] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 816.843310] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 816.843310] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 88d10958-f4d5-4378-98b5-dae42820a117, please check neutron logs for more information. [ 816.843310] env[61906]: ERROR nova.compute.manager [ 816.843310] env[61906]: Traceback (most recent call last): [ 816.843447] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 816.843447] env[61906]: listener.cb(fileno) [ 816.843447] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 816.843447] env[61906]: result = function(*args, **kwargs) [ 816.843447] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 816.843447] env[61906]: return func(*args, **kwargs) [ 816.843447] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 816.843447] env[61906]: raise e [ 816.843447] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 816.843447] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 816.843447] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 816.843447] env[61906]: created_port_ids = self._update_ports_for_instance( [ 816.843447] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 816.843447] env[61906]: with excutils.save_and_reraise_exception(): [ 816.843447] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 816.843447] env[61906]: self.force_reraise() [ 816.843447] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 816.843447] env[61906]: raise self.value [ 816.843447] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 816.843447] env[61906]: updated_port = self._update_port( [ 816.843447] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 816.843447] env[61906]: _ensure_no_port_binding_failure(port) [ 816.843447] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 816.843447] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 816.843447] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 88d10958-f4d5-4378-98b5-dae42820a117, please check neutron logs for more information. [ 816.844203] env[61906]: Removing descriptor: 18 [ 816.851030] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7af81fd8-50b8-4fab-bc09-5ceb0aebf209 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.868632] env[61906]: ERROR nova.compute.manager [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 88d10958-f4d5-4378-98b5-dae42820a117, please check neutron logs for more information. [ 816.868632] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Traceback (most recent call last): [ 816.868632] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 816.868632] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] yield resources [ 816.868632] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 816.868632] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] self.driver.spawn(context, instance, image_meta, [ 816.868632] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 816.868632] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 816.868632] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 816.868632] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] vm_ref = self.build_virtual_machine(instance, [ 816.868632] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 816.868940] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] vif_infos = vmwarevif.get_vif_info(self._session, [ 816.868940] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 816.868940] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] for vif in network_info: [ 816.868940] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 816.868940] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] return self._sync_wrapper(fn, *args, **kwargs) [ 816.868940] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 816.868940] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] self.wait() [ 816.868940] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 816.868940] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] self[:] = self._gt.wait() [ 816.868940] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 816.868940] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] return self._exit_event.wait() [ 816.868940] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 816.868940] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] current.throw(*self._exc) [ 816.869241] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 816.869241] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] result = function(*args, **kwargs) [ 816.869241] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 816.869241] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] return func(*args, **kwargs) [ 816.869241] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 816.869241] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] raise e [ 816.869241] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 816.869241] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] nwinfo = self.network_api.allocate_for_instance( [ 816.869241] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 816.869241] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] created_port_ids = self._update_ports_for_instance( [ 816.869241] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 816.869241] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] with excutils.save_and_reraise_exception(): [ 816.869241] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 816.869642] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] self.force_reraise() [ 816.869642] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 816.869642] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] raise self.value [ 816.869642] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 816.869642] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] updated_port = self._update_port( [ 816.869642] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 816.869642] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] _ensure_no_port_binding_failure(port) [ 816.869642] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 816.869642] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] raise exception.PortBindingFailed(port_id=port['id']) [ 816.869642] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] nova.exception.PortBindingFailed: Binding failed for port 88d10958-f4d5-4378-98b5-dae42820a117, please check neutron logs for more information. [ 816.869642] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] [ 816.869642] env[61906]: INFO nova.compute.manager [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Terminating instance [ 816.873445] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Acquiring lock "refresh_cache-3923d04e-5b50-4cfc-b3da-4b4eada4c8e3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.873445] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Acquired lock "refresh_cache-3923d04e-5b50-4cfc-b3da-4b4eada4c8e3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.873445] env[61906]: DEBUG nova.network.neutron [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 816.939553] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52946a5e-a09c-7b0c-3c88-4398e205d69b, 'name': SearchDatastore_Task, 'duration_secs': 0.00908} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.939842] env[61906]: DEBUG oslo_concurrency.lockutils [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.940614] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 816.940614] env[61906]: DEBUG oslo_concurrency.lockutils [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.940614] env[61906]: DEBUG oslo_concurrency.lockutils [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.940614] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 816.940872] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-676f19b1-1fd5-46e3-88d9-55e0beb79b13 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.949208] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 816.949305] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 816.950090] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5da64837-7f98-4fb2-9543-ba59f8b7083e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.956017] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Waiting for the task: (returnval){ [ 816.956017] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]524600cd-7a9d-7e60-9449-1c080e8a61da" [ 816.956017] env[61906]: _type = "Task" [ 816.956017] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.964564] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]524600cd-7a9d-7e60-9449-1c080e8a61da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.169735] env[61906]: DEBUG nova.network.neutron [-] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.216415] env[61906]: DEBUG oslo_concurrency.lockutils [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.434s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.216965] env[61906]: DEBUG nova.compute.manager [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 817.220273] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.061s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.405525] env[61906]: DEBUG nova.network.neutron [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 817.470981] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]524600cd-7a9d-7e60-9449-1c080e8a61da, 'name': SearchDatastore_Task, 'duration_secs': 0.00915} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.472282] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c2d63d7-8ae9-4c1e-8a1f-294aa4ce7368 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.479751] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Waiting for the task: (returnval){ [ 817.479751] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52eedca1-e682-b9a1-8481-301943bf42ab" [ 817.479751] env[61906]: _type = "Task" [ 817.479751] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.491206] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52eedca1-e682-b9a1-8481-301943bf42ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.505961] env[61906]: DEBUG nova.network.neutron [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.671968] env[61906]: INFO nova.compute.manager [-] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Took 1.03 seconds to deallocate network for instance. [ 817.677019] env[61906]: DEBUG nova.compute.claims [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 817.677019] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.721970] env[61906]: DEBUG nova.compute.utils [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 817.723399] env[61906]: DEBUG nova.compute.manager [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 817.723573] env[61906]: DEBUG nova.network.neutron [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 817.794356] env[61906]: DEBUG nova.policy [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f1b4e2421d4a40e399670e550bf5f011', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '94ad46a8089a446181abde0f6361b343', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 817.861267] env[61906]: DEBUG nova.compute.manager [req-63b75f9d-a3c4-443a-bf2f-63673810e72d req-4149aa72-fd25-448e-99c3-2c5fa8e6addb service nova] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Received event network-changed-88d10958-f4d5-4378-98b5-dae42820a117 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 817.861489] env[61906]: DEBUG nova.compute.manager [req-63b75f9d-a3c4-443a-bf2f-63673810e72d req-4149aa72-fd25-448e-99c3-2c5fa8e6addb service nova] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Refreshing instance network info cache due to event network-changed-88d10958-f4d5-4378-98b5-dae42820a117. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 817.861785] env[61906]: DEBUG oslo_concurrency.lockutils [req-63b75f9d-a3c4-443a-bf2f-63673810e72d req-4149aa72-fd25-448e-99c3-2c5fa8e6addb service nova] Acquiring lock "refresh_cache-3923d04e-5b50-4cfc-b3da-4b4eada4c8e3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.991257] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52eedca1-e682-b9a1-8481-301943bf42ab, 'name': SearchDatastore_Task, 'duration_secs': 0.009356} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.993923] env[61906]: DEBUG oslo_concurrency.lockutils [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.994710] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 7e452659-0a5e-4740-b7ed-1fd7b1917dac/7e452659-0a5e-4740-b7ed-1fd7b1917dac.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 817.996145] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b60d8d1c-bd93-4444-b44f-353715cc4301 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.003736] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Waiting for the task: (returnval){ [ 818.003736] env[61906]: value = "task-1333154" [ 818.003736] env[61906]: _type = "Task" [ 818.003736] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.013643] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Releasing lock "refresh_cache-3923d04e-5b50-4cfc-b3da-4b4eada4c8e3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.014217] env[61906]: DEBUG nova.compute.manager [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 818.015404] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 818.016159] env[61906]: DEBUG oslo_concurrency.lockutils [req-63b75f9d-a3c4-443a-bf2f-63673810e72d req-4149aa72-fd25-448e-99c3-2c5fa8e6addb service nova] Acquired lock "refresh_cache-3923d04e-5b50-4cfc-b3da-4b4eada4c8e3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.019081] env[61906]: DEBUG nova.network.neutron [req-63b75f9d-a3c4-443a-bf2f-63673810e72d req-4149aa72-fd25-448e-99c3-2c5fa8e6addb service nova] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Refreshing network info cache for port 88d10958-f4d5-4378-98b5-dae42820a117 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 818.019081] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-314e8eb3-a18d-42f5-ba60-97da09a0a32a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.027240] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333154, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.030985] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55669153-ba60-4146-9c51-33a08543f300 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.038696] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-767da228-d782-4b58-9bcd-bbf18edb928d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.053774] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0584fba-797c-443d-bc1e-79bf0a0259eb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.065251] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3 could not be found. [ 818.065251] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 818.065251] env[61906]: INFO nova.compute.manager [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Took 0.05 seconds to destroy the instance on the hypervisor. [ 818.065251] env[61906]: DEBUG oslo.service.loopingcall [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 818.065251] env[61906]: DEBUG nova.compute.manager [-] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 818.065251] env[61906]: DEBUG nova.network.neutron [-] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 818.095425] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-412e73cd-6503-4ca2-b2bc-aa4c2bf17725 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.103853] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17e44667-7e9f-4ef0-a4ae-07d70440292e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.118947] env[61906]: DEBUG nova.compute.provider_tree [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.124021] env[61906]: DEBUG nova.network.neutron [-] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 818.135058] env[61906]: DEBUG nova.network.neutron [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Successfully created port: 5a9efc88-6a93-4307-859b-8286ddc55f73 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 818.226968] env[61906]: DEBUG nova.compute.manager [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 818.515115] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333154, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485262} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.515433] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 7e452659-0a5e-4740-b7ed-1fd7b1917dac/7e452659-0a5e-4740-b7ed-1fd7b1917dac.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 818.516058] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 818.516439] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-96ca2075-2191-4759-8c91-6d672757eb40 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.526498] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Waiting for the task: (returnval){ [ 818.526498] env[61906]: value = "task-1333155" [ 818.526498] env[61906]: _type = "Task" [ 818.526498] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.540728] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333155, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.546945] env[61906]: DEBUG nova.network.neutron [req-63b75f9d-a3c4-443a-bf2f-63673810e72d req-4149aa72-fd25-448e-99c3-2c5fa8e6addb service nova] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 818.623475] env[61906]: DEBUG nova.scheduler.client.report [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 818.627030] env[61906]: DEBUG nova.network.neutron [-] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.694972] env[61906]: DEBUG nova.network.neutron [req-63b75f9d-a3c4-443a-bf2f-63673810e72d req-4149aa72-fd25-448e-99c3-2c5fa8e6addb service nova] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.037808] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333155, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058092} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.038232] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 819.039062] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8579c43-256f-4710-8ea6-903f4ac59873 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.059657] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] 7e452659-0a5e-4740-b7ed-1fd7b1917dac/7e452659-0a5e-4740-b7ed-1fd7b1917dac.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 819.061616] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-54518e4b-c50d-4840-a737-40dd6be2a934 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.081236] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Waiting for the task: (returnval){ [ 819.081236] env[61906]: value = "task-1333156" [ 819.081236] env[61906]: _type = "Task" [ 819.081236] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.089589] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333156, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.129926] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.909s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.130594] env[61906]: ERROR nova.compute.manager [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b01c48b8-7698-4df9-8653-500e06a0f61a, please check neutron logs for more information. [ 819.130594] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Traceback (most recent call last): [ 819.130594] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 819.130594] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] self.driver.spawn(context, instance, image_meta, [ 819.130594] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 819.130594] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 819.130594] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 819.130594] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] vm_ref = self.build_virtual_machine(instance, [ 819.130594] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 819.130594] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] vif_infos = vmwarevif.get_vif_info(self._session, [ 819.130594] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 819.130912] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] for vif in network_info: [ 819.130912] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 819.130912] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] return self._sync_wrapper(fn, *args, **kwargs) [ 819.130912] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 819.130912] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] self.wait() [ 819.130912] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 819.130912] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] self[:] = self._gt.wait() [ 819.130912] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 819.130912] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] return self._exit_event.wait() [ 819.130912] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 819.130912] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] result = hub.switch() [ 819.130912] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 819.130912] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] return self.greenlet.switch() [ 819.131275] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 819.131275] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] result = function(*args, **kwargs) [ 819.131275] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 819.131275] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] return func(*args, **kwargs) [ 819.131275] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 819.131275] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] raise e [ 819.131275] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 819.131275] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] nwinfo = self.network_api.allocate_for_instance( [ 819.131275] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 819.131275] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] created_port_ids = self._update_ports_for_instance( [ 819.131275] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 819.131275] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] with excutils.save_and_reraise_exception(): [ 819.131275] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 819.131598] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] self.force_reraise() [ 819.131598] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 819.131598] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] raise self.value [ 819.131598] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 819.131598] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] updated_port = self._update_port( [ 819.131598] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 819.131598] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] _ensure_no_port_binding_failure(port) [ 819.131598] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 819.131598] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] raise exception.PortBindingFailed(port_id=port['id']) [ 819.131598] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] nova.exception.PortBindingFailed: Binding failed for port b01c48b8-7698-4df9-8653-500e06a0f61a, please check neutron logs for more information. [ 819.131598] env[61906]: ERROR nova.compute.manager [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] [ 819.131918] env[61906]: DEBUG nova.compute.utils [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Binding failed for port b01c48b8-7698-4df9-8653-500e06a0f61a, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 819.133042] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.224s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.133452] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.133685] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61906) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 819.134019] env[61906]: DEBUG oslo_concurrency.lockutils [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.985s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.136327] env[61906]: INFO nova.compute.claims [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 819.139452] env[61906]: INFO nova.compute.manager [-] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Took 1.08 seconds to deallocate network for instance. [ 819.139965] env[61906]: DEBUG nova.compute.manager [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Build of instance 7ad136a2-d850-4b68-af0f-73badf65426c was re-scheduled: Binding failed for port b01c48b8-7698-4df9-8653-500e06a0f61a, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 819.140588] env[61906]: DEBUG nova.compute.manager [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 819.140866] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Acquiring lock "refresh_cache-7ad136a2-d850-4b68-af0f-73badf65426c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.141054] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Acquired lock "refresh_cache-7ad136a2-d850-4b68-af0f-73badf65426c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.141249] env[61906]: DEBUG nova.network.neutron [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 819.142879] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01b51c75-ca3c-4c37-b47b-d7bb685a4a42 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.151672] env[61906]: DEBUG nova.compute.claims [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 819.152350] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.158730] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-127f8824-870c-474d-b5dd-c7962f08e4cf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.175936] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e107447-a08f-4cfc-8cb8-120277089630 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.185126] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10250e16-8a90-4d40-af8b-35fb78278646 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.218172] env[61906]: DEBUG oslo_concurrency.lockutils [req-63b75f9d-a3c4-443a-bf2f-63673810e72d req-4149aa72-fd25-448e-99c3-2c5fa8e6addb service nova] Releasing lock "refresh_cache-3923d04e-5b50-4cfc-b3da-4b4eada4c8e3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.218172] env[61906]: DEBUG nova.compute.manager [req-63b75f9d-a3c4-443a-bf2f-63673810e72d req-4149aa72-fd25-448e-99c3-2c5fa8e6addb service nova] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Received event network-vif-deleted-88d10958-f4d5-4378-98b5-dae42820a117 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 819.218891] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181430MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=61906) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 819.218891] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.241333] env[61906]: DEBUG nova.compute.manager [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 819.267255] env[61906]: DEBUG nova.virt.hardware [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 819.267689] env[61906]: DEBUG nova.virt.hardware [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 819.267689] env[61906]: DEBUG nova.virt.hardware [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 819.267885] env[61906]: DEBUG nova.virt.hardware [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 819.268105] env[61906]: DEBUG nova.virt.hardware [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 819.268366] env[61906]: DEBUG nova.virt.hardware [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 819.268596] env[61906]: DEBUG nova.virt.hardware [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 819.268784] env[61906]: DEBUG nova.virt.hardware [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 819.269128] env[61906]: DEBUG nova.virt.hardware [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 819.269341] env[61906]: DEBUG nova.virt.hardware [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 819.269743] env[61906]: DEBUG nova.virt.hardware [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 819.270492] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-858894e2-b642-4b69-ad61-907e8e8b0f2c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.274125] env[61906]: ERROR nova.compute.manager [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5a9efc88-6a93-4307-859b-8286ddc55f73, please check neutron logs for more information. [ 819.274125] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 819.274125] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 819.274125] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 819.274125] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 819.274125] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 819.274125] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 819.274125] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 819.274125] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 819.274125] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 819.274125] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 819.274125] env[61906]: ERROR nova.compute.manager raise self.value [ 819.274125] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 819.274125] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 819.274125] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 819.274125] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 819.275277] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 819.275277] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 819.275277] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5a9efc88-6a93-4307-859b-8286ddc55f73, please check neutron logs for more information. [ 819.275277] env[61906]: ERROR nova.compute.manager [ 819.275277] env[61906]: Traceback (most recent call last): [ 819.275277] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 819.275277] env[61906]: listener.cb(fileno) [ 819.275277] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 819.275277] env[61906]: result = function(*args, **kwargs) [ 819.275277] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 819.275277] env[61906]: return func(*args, **kwargs) [ 819.275277] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 819.275277] env[61906]: raise e [ 819.275277] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 819.275277] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 819.275277] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 819.275277] env[61906]: created_port_ids = self._update_ports_for_instance( [ 819.275277] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 819.275277] env[61906]: with excutils.save_and_reraise_exception(): [ 819.275277] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 819.275277] env[61906]: self.force_reraise() [ 819.275277] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 819.275277] env[61906]: raise self.value [ 819.275277] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 819.275277] env[61906]: updated_port = self._update_port( [ 819.275277] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 819.275277] env[61906]: _ensure_no_port_binding_failure(port) [ 819.275277] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 819.275277] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 819.276011] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 5a9efc88-6a93-4307-859b-8286ddc55f73, please check neutron logs for more information. [ 819.276011] env[61906]: Removing descriptor: 18 [ 819.280652] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1786ec77-c1d0-406a-93c2-7fcc01b61fcc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.295863] env[61906]: ERROR nova.compute.manager [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5a9efc88-6a93-4307-859b-8286ddc55f73, please check neutron logs for more information. [ 819.295863] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Traceback (most recent call last): [ 819.295863] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 819.295863] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] yield resources [ 819.295863] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 819.295863] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] self.driver.spawn(context, instance, image_meta, [ 819.295863] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 819.295863] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] self._vmops.spawn(context, instance, image_meta, injected_files, [ 819.295863] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 819.295863] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] vm_ref = self.build_virtual_machine(instance, [ 819.295863] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 819.296247] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] vif_infos = vmwarevif.get_vif_info(self._session, [ 819.296247] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 819.296247] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] for vif in network_info: [ 819.296247] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 819.296247] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] return self._sync_wrapper(fn, *args, **kwargs) [ 819.296247] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 819.296247] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] self.wait() [ 819.296247] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 819.296247] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] self[:] = self._gt.wait() [ 819.296247] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 819.296247] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] return self._exit_event.wait() [ 819.296247] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 819.296247] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] current.throw(*self._exc) [ 819.296544] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 819.296544] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] result = function(*args, **kwargs) [ 819.296544] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 819.296544] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] return func(*args, **kwargs) [ 819.296544] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 819.296544] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] raise e [ 819.296544] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 819.296544] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] nwinfo = self.network_api.allocate_for_instance( [ 819.296544] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 819.296544] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] created_port_ids = self._update_ports_for_instance( [ 819.296544] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 819.296544] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] with excutils.save_and_reraise_exception(): [ 819.296544] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 819.296928] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] self.force_reraise() [ 819.296928] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 819.296928] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] raise self.value [ 819.296928] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 819.296928] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] updated_port = self._update_port( [ 819.296928] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 819.296928] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] _ensure_no_port_binding_failure(port) [ 819.296928] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 819.296928] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] raise exception.PortBindingFailed(port_id=port['id']) [ 819.296928] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] nova.exception.PortBindingFailed: Binding failed for port 5a9efc88-6a93-4307-859b-8286ddc55f73, please check neutron logs for more information. [ 819.296928] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] [ 819.296928] env[61906]: INFO nova.compute.manager [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Terminating instance [ 819.298560] env[61906]: DEBUG oslo_concurrency.lockutils [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Acquiring lock "refresh_cache-df9c9d62-8dc5-452c-85ed-846dcbd76014" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.298718] env[61906]: DEBUG oslo_concurrency.lockutils [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Acquired lock "refresh_cache-df9c9d62-8dc5-452c-85ed-846dcbd76014" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.298931] env[61906]: DEBUG nova.network.neutron [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 819.591808] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333156, 'name': ReconfigVM_Task, 'duration_secs': 0.274316} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.592103] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Reconfigured VM instance instance-00000033 to attach disk [datastore2] 7e452659-0a5e-4740-b7ed-1fd7b1917dac/7e452659-0a5e-4740-b7ed-1fd7b1917dac.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 819.592744] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1033ae81-69b8-4ad4-bd8d-30d24ae4d1d1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.600980] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Waiting for the task: (returnval){ [ 819.600980] env[61906]: value = "task-1333157" [ 819.600980] env[61906]: _type = "Task" [ 819.600980] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.610502] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333157, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.669408] env[61906]: DEBUG nova.network.neutron [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 819.766958] env[61906]: DEBUG nova.network.neutron [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.822858] env[61906]: DEBUG nova.network.neutron [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 819.890849] env[61906]: DEBUG nova.compute.manager [req-b91766e6-85b7-4b38-80aa-6f69f5e3ec8b req-71fe4c31-25f8-44a3-b130-88a63497d831 service nova] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Received event network-changed-5a9efc88-6a93-4307-859b-8286ddc55f73 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 819.891069] env[61906]: DEBUG nova.compute.manager [req-b91766e6-85b7-4b38-80aa-6f69f5e3ec8b req-71fe4c31-25f8-44a3-b130-88a63497d831 service nova] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Refreshing instance network info cache due to event network-changed-5a9efc88-6a93-4307-859b-8286ddc55f73. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 819.891262] env[61906]: DEBUG oslo_concurrency.lockutils [req-b91766e6-85b7-4b38-80aa-6f69f5e3ec8b req-71fe4c31-25f8-44a3-b130-88a63497d831 service nova] Acquiring lock "refresh_cache-df9c9d62-8dc5-452c-85ed-846dcbd76014" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.923117] env[61906]: DEBUG nova.network.neutron [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.111306] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333157, 'name': Rename_Task, 'duration_secs': 0.146076} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.111800] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 820.112076] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e2429c86-93e2-4119-8ba2-766617a5a1b6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.119132] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Waiting for the task: (returnval){ [ 820.119132] env[61906]: value = "task-1333159" [ 820.119132] env[61906]: _type = "Task" [ 820.119132] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.127600] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333159, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.272868] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Releasing lock "refresh_cache-7ad136a2-d850-4b68-af0f-73badf65426c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.273161] env[61906]: DEBUG nova.compute.manager [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 820.273426] env[61906]: DEBUG nova.compute.manager [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 820.274054] env[61906]: DEBUG nova.network.neutron [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 820.290230] env[61906]: DEBUG nova.network.neutron [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 820.408870] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e4b6857-86c5-4b14-a4b5-9d9433f90edd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.417150] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ee1c984-ac88-4ce5-ac24-f1a5646152ff {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.447765] env[61906]: DEBUG oslo_concurrency.lockutils [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Releasing lock "refresh_cache-df9c9d62-8dc5-452c-85ed-846dcbd76014" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.453018] env[61906]: DEBUG nova.compute.manager [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 820.453018] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 820.453018] env[61906]: DEBUG oslo_concurrency.lockutils [req-b91766e6-85b7-4b38-80aa-6f69f5e3ec8b req-71fe4c31-25f8-44a3-b130-88a63497d831 service nova] Acquired lock "refresh_cache-df9c9d62-8dc5-452c-85ed-846dcbd76014" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.453018] env[61906]: DEBUG nova.network.neutron [req-b91766e6-85b7-4b38-80aa-6f69f5e3ec8b req-71fe4c31-25f8-44a3-b130-88a63497d831 service nova] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Refreshing network info cache for port 5a9efc88-6a93-4307-859b-8286ddc55f73 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 820.453018] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5f52290c-d820-4b6e-b0fb-9b37d7b6f712 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.454656] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f65cce93-9364-4481-8ddc-3751fe31e12f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.464565] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2bc2933-472f-40fb-a051-45468b9fc202 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.472633] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d7d182d-8ef0-4452-8008-4075b49365c8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.497386] env[61906]: DEBUG nova.compute.provider_tree [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 820.503358] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance df9c9d62-8dc5-452c-85ed-846dcbd76014 could not be found. [ 820.503921] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 820.504492] env[61906]: INFO nova.compute.manager [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Took 0.05 seconds to destroy the instance on the hypervisor. [ 820.504819] env[61906]: DEBUG oslo.service.loopingcall [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 820.505310] env[61906]: DEBUG nova.compute.manager [-] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 820.505467] env[61906]: DEBUG nova.network.neutron [-] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 820.522133] env[61906]: DEBUG nova.network.neutron [-] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 820.629498] env[61906]: DEBUG oslo_vmware.api [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333159, 'name': PowerOnVM_Task, 'duration_secs': 0.409469} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.629786] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 820.629983] env[61906]: DEBUG nova.compute.manager [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 820.630742] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25a83576-9b84-44e6-b04d-ad92b090f304 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.796445] env[61906]: DEBUG nova.network.neutron [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.011794] env[61906]: DEBUG nova.scheduler.client.report [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 821.025465] env[61906]: DEBUG nova.network.neutron [-] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.159542] env[61906]: DEBUG oslo_concurrency.lockutils [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.164403] env[61906]: DEBUG nova.network.neutron [req-b91766e6-85b7-4b38-80aa-6f69f5e3ec8b req-71fe4c31-25f8-44a3-b130-88a63497d831 service nova] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 821.262941] env[61906]: DEBUG nova.network.neutron [req-b91766e6-85b7-4b38-80aa-6f69f5e3ec8b req-71fe4c31-25f8-44a3-b130-88a63497d831 service nova] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.299920] env[61906]: INFO nova.compute.manager [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] [instance: 7ad136a2-d850-4b68-af0f-73badf65426c] Took 1.03 seconds to deallocate network for instance. [ 821.519024] env[61906]: DEBUG oslo_concurrency.lockutils [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.384s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.519024] env[61906]: DEBUG nova.compute.manager [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 821.521168] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.551s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.523027] env[61906]: INFO nova.compute.claims [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 821.526815] env[61906]: INFO nova.compute.manager [-] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Took 1.02 seconds to deallocate network for instance. [ 821.528895] env[61906]: DEBUG nova.compute.claims [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 821.528977] env[61906]: DEBUG oslo_concurrency.lockutils [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.768961] env[61906]: DEBUG oslo_concurrency.lockutils [req-b91766e6-85b7-4b38-80aa-6f69f5e3ec8b req-71fe4c31-25f8-44a3-b130-88a63497d831 service nova] Releasing lock "refresh_cache-df9c9d62-8dc5-452c-85ed-846dcbd76014" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.768961] env[61906]: DEBUG nova.compute.manager [req-b91766e6-85b7-4b38-80aa-6f69f5e3ec8b req-71fe4c31-25f8-44a3-b130-88a63497d831 service nova] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Received event network-vif-deleted-5a9efc88-6a93-4307-859b-8286ddc55f73 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.027490] env[61906]: DEBUG nova.compute.utils [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 822.031042] env[61906]: DEBUG nova.compute.manager [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 822.031222] env[61906]: DEBUG nova.network.neutron [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 822.071933] env[61906]: DEBUG nova.policy [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ff0b66e5d68b402199d6c86ad92e2db7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8226b66484094affad3e16d2ab04accd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 822.338256] env[61906]: INFO nova.scheduler.client.report [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Deleted allocations for instance 7ad136a2-d850-4b68-af0f-73badf65426c [ 822.344147] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Acquiring lock "7e452659-0a5e-4740-b7ed-1fd7b1917dac" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.344405] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Lock "7e452659-0a5e-4740-b7ed-1fd7b1917dac" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.344980] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Acquiring lock "7e452659-0a5e-4740-b7ed-1fd7b1917dac-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.345175] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Lock "7e452659-0a5e-4740-b7ed-1fd7b1917dac-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.345353] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Lock "7e452659-0a5e-4740-b7ed-1fd7b1917dac-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.347985] env[61906]: INFO nova.compute.manager [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Terminating instance [ 822.349754] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Acquiring lock "refresh_cache-7e452659-0a5e-4740-b7ed-1fd7b1917dac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.349900] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Acquired lock "refresh_cache-7e452659-0a5e-4740-b7ed-1fd7b1917dac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.350137] env[61906]: DEBUG nova.network.neutron [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 822.393317] env[61906]: DEBUG nova.network.neutron [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Successfully created port: 018c3cd4-50ad-4134-8ee4-c8aff36ab617 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 822.416458] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "cc8870d0-eafb-4f73-aa79-b98f51370237" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.416714] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "cc8870d0-eafb-4f73-aa79-b98f51370237" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.532024] env[61906]: DEBUG nova.compute.manager [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 822.828433] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c202e526-6a45-4113-aa9b-580b70047b18 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.838606] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-931ce843-0719-41f5-a11e-e03afb64dc24 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.879716] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b3b8ec41-f35c-403f-b8af-b70269231a73 tempest-ServersV294TestFqdnHostnames-1041166524 tempest-ServersV294TestFqdnHostnames-1041166524-project-member] Lock "7ad136a2-d850-4b68-af0f-73badf65426c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 176.119s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.883747] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68e82998-7100-409b-9344-7b020934c490 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.892985] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb88988a-c0b2-441b-9e22-0b102204ba9c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.908466] env[61906]: DEBUG nova.compute.provider_tree [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.910617] env[61906]: DEBUG nova.network.neutron [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 823.014318] env[61906]: DEBUG nova.network.neutron [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.386357] env[61906]: DEBUG nova.compute.manager [req-3e98fdd0-b6d1-464a-8372-02b5fb5bdf61 req-5fc32855-a70a-4459-85d1-67439206520e service nova] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Received event network-changed-018c3cd4-50ad-4134-8ee4-c8aff36ab617 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 823.386357] env[61906]: DEBUG nova.compute.manager [req-3e98fdd0-b6d1-464a-8372-02b5fb5bdf61 req-5fc32855-a70a-4459-85d1-67439206520e service nova] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Refreshing instance network info cache due to event network-changed-018c3cd4-50ad-4134-8ee4-c8aff36ab617. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 823.386357] env[61906]: DEBUG oslo_concurrency.lockutils [req-3e98fdd0-b6d1-464a-8372-02b5fb5bdf61 req-5fc32855-a70a-4459-85d1-67439206520e service nova] Acquiring lock "refresh_cache-52a387d2-4090-4eba-9a77-98c37f152b75" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.386357] env[61906]: DEBUG oslo_concurrency.lockutils [req-3e98fdd0-b6d1-464a-8372-02b5fb5bdf61 req-5fc32855-a70a-4459-85d1-67439206520e service nova] Acquired lock "refresh_cache-52a387d2-4090-4eba-9a77-98c37f152b75" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.386357] env[61906]: DEBUG nova.network.neutron [req-3e98fdd0-b6d1-464a-8372-02b5fb5bdf61 req-5fc32855-a70a-4459-85d1-67439206520e service nova] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Refreshing network info cache for port 018c3cd4-50ad-4134-8ee4-c8aff36ab617 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 823.389430] env[61906]: DEBUG nova.compute.manager [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 823.416036] env[61906]: DEBUG nova.scheduler.client.report [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 823.517453] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Releasing lock "refresh_cache-7e452659-0a5e-4740-b7ed-1fd7b1917dac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.518025] env[61906]: DEBUG nova.compute.manager [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 823.518353] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 823.519576] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d1c70ef-8038-4e53-9bd1-a42bbb14010e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.530435] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 823.531494] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f494525e-9b4b-4ac3-b1e1-c57350983e04 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.538336] env[61906]: DEBUG oslo_vmware.api [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Waiting for the task: (returnval){ [ 823.538336] env[61906]: value = "task-1333161" [ 823.538336] env[61906]: _type = "Task" [ 823.538336] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.547641] env[61906]: DEBUG nova.compute.manager [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 823.555187] env[61906]: DEBUG oslo_vmware.api [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333161, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.574121] env[61906]: DEBUG nova.virt.hardware [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 823.574361] env[61906]: DEBUG nova.virt.hardware [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 823.574512] env[61906]: DEBUG nova.virt.hardware [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 823.574699] env[61906]: DEBUG nova.virt.hardware [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 823.574840] env[61906]: DEBUG nova.virt.hardware [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 823.574983] env[61906]: DEBUG nova.virt.hardware [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 823.575510] env[61906]: DEBUG nova.virt.hardware [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 823.575740] env[61906]: DEBUG nova.virt.hardware [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 823.575965] env[61906]: DEBUG nova.virt.hardware [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 823.576210] env[61906]: DEBUG nova.virt.hardware [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 823.576443] env[61906]: DEBUG nova.virt.hardware [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 823.577619] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e82ebc5a-5388-4117-affc-ead794b6ca9a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.588282] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1012af5a-4893-4420-be24-db52def99ab5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.610501] env[61906]: ERROR nova.compute.manager [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 018c3cd4-50ad-4134-8ee4-c8aff36ab617, please check neutron logs for more information. [ 823.610501] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 823.610501] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 823.610501] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 823.610501] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 823.610501] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 823.610501] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 823.610501] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 823.610501] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 823.610501] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 823.610501] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 823.610501] env[61906]: ERROR nova.compute.manager raise self.value [ 823.610501] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 823.610501] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 823.610501] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 823.610501] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 823.610922] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 823.610922] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 823.610922] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 018c3cd4-50ad-4134-8ee4-c8aff36ab617, please check neutron logs for more information. [ 823.610922] env[61906]: ERROR nova.compute.manager [ 823.610922] env[61906]: Traceback (most recent call last): [ 823.610922] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 823.610922] env[61906]: listener.cb(fileno) [ 823.610922] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 823.610922] env[61906]: result = function(*args, **kwargs) [ 823.610922] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 823.610922] env[61906]: return func(*args, **kwargs) [ 823.610922] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 823.610922] env[61906]: raise e [ 823.610922] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 823.610922] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 823.610922] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 823.610922] env[61906]: created_port_ids = self._update_ports_for_instance( [ 823.610922] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 823.610922] env[61906]: with excutils.save_and_reraise_exception(): [ 823.610922] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 823.610922] env[61906]: self.force_reraise() [ 823.610922] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 823.610922] env[61906]: raise self.value [ 823.610922] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 823.610922] env[61906]: updated_port = self._update_port( [ 823.610922] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 823.610922] env[61906]: _ensure_no_port_binding_failure(port) [ 823.610922] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 823.610922] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 823.611649] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 018c3cd4-50ad-4134-8ee4-c8aff36ab617, please check neutron logs for more information. [ 823.611649] env[61906]: Removing descriptor: 18 [ 823.611649] env[61906]: ERROR nova.compute.manager [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 018c3cd4-50ad-4134-8ee4-c8aff36ab617, please check neutron logs for more information. [ 823.611649] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Traceback (most recent call last): [ 823.611649] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 823.611649] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] yield resources [ 823.611649] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 823.611649] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] self.driver.spawn(context, instance, image_meta, [ 823.611649] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 823.611649] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] self._vmops.spawn(context, instance, image_meta, injected_files, [ 823.611649] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 823.611649] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] vm_ref = self.build_virtual_machine(instance, [ 823.611929] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 823.611929] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] vif_infos = vmwarevif.get_vif_info(self._session, [ 823.611929] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 823.611929] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] for vif in network_info: [ 823.611929] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 823.611929] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] return self._sync_wrapper(fn, *args, **kwargs) [ 823.611929] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 823.611929] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] self.wait() [ 823.611929] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 823.611929] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] self[:] = self._gt.wait() [ 823.611929] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 823.611929] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] return self._exit_event.wait() [ 823.611929] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 823.612228] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] result = hub.switch() [ 823.612228] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 823.612228] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] return self.greenlet.switch() [ 823.612228] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 823.612228] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] result = function(*args, **kwargs) [ 823.612228] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 823.612228] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] return func(*args, **kwargs) [ 823.612228] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 823.612228] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] raise e [ 823.612228] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 823.612228] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] nwinfo = self.network_api.allocate_for_instance( [ 823.612228] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 823.612228] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] created_port_ids = self._update_ports_for_instance( [ 823.612561] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 823.612561] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] with excutils.save_and_reraise_exception(): [ 823.612561] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 823.612561] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] self.force_reraise() [ 823.612561] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 823.612561] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] raise self.value [ 823.612561] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 823.612561] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] updated_port = self._update_port( [ 823.612561] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 823.612561] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] _ensure_no_port_binding_failure(port) [ 823.612561] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 823.612561] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] raise exception.PortBindingFailed(port_id=port['id']) [ 823.612898] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] nova.exception.PortBindingFailed: Binding failed for port 018c3cd4-50ad-4134-8ee4-c8aff36ab617, please check neutron logs for more information. [ 823.612898] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] [ 823.612898] env[61906]: INFO nova.compute.manager [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Terminating instance [ 823.616047] env[61906]: DEBUG oslo_concurrency.lockutils [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Acquiring lock "refresh_cache-52a387d2-4090-4eba-9a77-98c37f152b75" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.916949] env[61906]: DEBUG nova.network.neutron [req-3e98fdd0-b6d1-464a-8372-02b5fb5bdf61 req-5fc32855-a70a-4459-85d1-67439206520e service nova] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 823.916949] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.923707] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.400s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.923707] env[61906]: DEBUG nova.compute.manager [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 823.924677] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.758s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.925454] env[61906]: INFO nova.compute.claims [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 824.038238] env[61906]: DEBUG nova.network.neutron [req-3e98fdd0-b6d1-464a-8372-02b5fb5bdf61 req-5fc32855-a70a-4459-85d1-67439206520e service nova] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.050974] env[61906]: DEBUG oslo_vmware.api [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333161, 'name': PowerOffVM_Task, 'duration_secs': 0.158708} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.054962] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 824.054962] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 824.054962] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ed6cf7a4-4843-45f6-9cad-d4ac786b195d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.083544] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 824.084256] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 824.084552] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Deleting the datastore file [datastore2] 7e452659-0a5e-4740-b7ed-1fd7b1917dac {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 824.084914] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-046e884f-b6d1-48b2-951b-5f96d92b7c98 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.096419] env[61906]: DEBUG oslo_vmware.api [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Waiting for the task: (returnval){ [ 824.096419] env[61906]: value = "task-1333164" [ 824.096419] env[61906]: _type = "Task" [ 824.096419] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.108521] env[61906]: DEBUG oslo_vmware.api [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333164, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.433074] env[61906]: DEBUG nova.compute.utils [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 824.437215] env[61906]: DEBUG nova.compute.manager [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 824.437300] env[61906]: DEBUG nova.network.neutron [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 824.519173] env[61906]: DEBUG nova.policy [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '684bd1e8957346379daafc095f66b83a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fc56ec8d118a4fe886829d946f647b00', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 824.545608] env[61906]: DEBUG oslo_concurrency.lockutils [req-3e98fdd0-b6d1-464a-8372-02b5fb5bdf61 req-5fc32855-a70a-4459-85d1-67439206520e service nova] Releasing lock "refresh_cache-52a387d2-4090-4eba-9a77-98c37f152b75" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.546034] env[61906]: DEBUG oslo_concurrency.lockutils [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Acquired lock "refresh_cache-52a387d2-4090-4eba-9a77-98c37f152b75" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.550019] env[61906]: DEBUG nova.network.neutron [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 824.610051] env[61906]: DEBUG oslo_vmware.api [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Task: {'id': task-1333164, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.106291} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.610302] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 824.610483] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 824.610664] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 824.610831] env[61906]: INFO nova.compute.manager [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Took 1.09 seconds to destroy the instance on the hypervisor. [ 824.611082] env[61906]: DEBUG oslo.service.loopingcall [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 824.611272] env[61906]: DEBUG nova.compute.manager [-] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 824.611363] env[61906]: DEBUG nova.network.neutron [-] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 824.645462] env[61906]: DEBUG nova.network.neutron [-] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 824.940018] env[61906]: DEBUG nova.compute.manager [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 824.976607] env[61906]: DEBUG nova.network.neutron [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Successfully created port: f16fa600-42c4-472d-8c4d-468806604d81 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 825.085393] env[61906]: DEBUG nova.network.neutron [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 825.148509] env[61906]: DEBUG nova.network.neutron [-] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.267664] env[61906]: DEBUG nova.network.neutron [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.303170] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a983f869-7de1-4e99-9561-b2c9e3279ed5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.311525] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bb6c778-c2be-4f98-aa79-bf002e814fb0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.343993] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34d42514-17f2-453b-bf31-73284888c1e7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.353026] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-192dd5ab-35bf-45a0-b319-4a42411d796d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.366945] env[61906]: DEBUG nova.compute.provider_tree [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 825.437726] env[61906]: DEBUG nova.compute.manager [req-3eb31e47-1810-4118-8b0d-fdf69e372b27 req-74741e15-e16e-48fa-83ad-0a39cb3d14a5 service nova] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Received event network-vif-deleted-018c3cd4-50ad-4134-8ee4-c8aff36ab617 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 825.653753] env[61906]: INFO nova.compute.manager [-] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Took 1.04 seconds to deallocate network for instance. [ 825.769851] env[61906]: DEBUG oslo_concurrency.lockutils [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Releasing lock "refresh_cache-52a387d2-4090-4eba-9a77-98c37f152b75" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.770607] env[61906]: DEBUG nova.compute.manager [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 825.770607] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 825.770804] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-be4c4c92-8852-4081-9f1d-bbc6af1ffdbb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.785241] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1237a9d2-1690-4ee2-b52a-0fb85f7bae4c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.810571] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 52a387d2-4090-4eba-9a77-98c37f152b75 could not be found. [ 825.810949] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 825.811259] env[61906]: INFO nova.compute.manager [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Took 0.04 seconds to destroy the instance on the hypervisor. [ 825.811525] env[61906]: DEBUG oslo.service.loopingcall [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 825.811831] env[61906]: DEBUG nova.compute.manager [-] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 825.811931] env[61906]: DEBUG nova.network.neutron [-] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 825.835375] env[61906]: DEBUG nova.network.neutron [-] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 825.869914] env[61906]: DEBUG nova.scheduler.client.report [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 825.951874] env[61906]: DEBUG nova.compute.manager [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 825.986476] env[61906]: DEBUG nova.virt.hardware [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 825.986735] env[61906]: DEBUG nova.virt.hardware [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 825.986894] env[61906]: DEBUG nova.virt.hardware [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 825.987089] env[61906]: DEBUG nova.virt.hardware [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 825.987240] env[61906]: DEBUG nova.virt.hardware [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 825.987388] env[61906]: DEBUG nova.virt.hardware [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 825.987592] env[61906]: DEBUG nova.virt.hardware [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 825.987754] env[61906]: DEBUG nova.virt.hardware [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 825.987973] env[61906]: DEBUG nova.virt.hardware [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 825.988309] env[61906]: DEBUG nova.virt.hardware [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 825.988634] env[61906]: DEBUG nova.virt.hardware [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 825.989668] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2828a939-46d9-4bdc-836b-0f932c55562f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.998504] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef042d53-92bf-4251-b39b-b1e6f9e3badf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.132754] env[61906]: ERROR nova.compute.manager [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f16fa600-42c4-472d-8c4d-468806604d81, please check neutron logs for more information. [ 826.132754] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 826.132754] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 826.132754] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 826.132754] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 826.132754] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 826.132754] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 826.132754] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 826.132754] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 826.132754] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 826.132754] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 826.132754] env[61906]: ERROR nova.compute.manager raise self.value [ 826.132754] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 826.132754] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 826.132754] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 826.132754] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 826.133241] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 826.133241] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 826.133241] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f16fa600-42c4-472d-8c4d-468806604d81, please check neutron logs for more information. [ 826.133241] env[61906]: ERROR nova.compute.manager [ 826.133241] env[61906]: Traceback (most recent call last): [ 826.133241] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 826.133241] env[61906]: listener.cb(fileno) [ 826.133241] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 826.133241] env[61906]: result = function(*args, **kwargs) [ 826.133241] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 826.133241] env[61906]: return func(*args, **kwargs) [ 826.133241] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 826.133241] env[61906]: raise e [ 826.133241] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 826.133241] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 826.133241] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 826.133241] env[61906]: created_port_ids = self._update_ports_for_instance( [ 826.133241] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 826.133241] env[61906]: with excutils.save_and_reraise_exception(): [ 826.133241] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 826.133241] env[61906]: self.force_reraise() [ 826.133241] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 826.133241] env[61906]: raise self.value [ 826.133241] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 826.133241] env[61906]: updated_port = self._update_port( [ 826.133241] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 826.133241] env[61906]: _ensure_no_port_binding_failure(port) [ 826.133241] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 826.133241] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 826.134861] env[61906]: nova.exception.PortBindingFailed: Binding failed for port f16fa600-42c4-472d-8c4d-468806604d81, please check neutron logs for more information. [ 826.134861] env[61906]: Removing descriptor: 18 [ 826.134861] env[61906]: ERROR nova.compute.manager [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f16fa600-42c4-472d-8c4d-468806604d81, please check neutron logs for more information. [ 826.134861] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Traceback (most recent call last): [ 826.134861] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 826.134861] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] yield resources [ 826.134861] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 826.134861] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] self.driver.spawn(context, instance, image_meta, [ 826.134861] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 826.134861] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] self._vmops.spawn(context, instance, image_meta, injected_files, [ 826.134861] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 826.134861] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] vm_ref = self.build_virtual_machine(instance, [ 826.135391] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 826.135391] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] vif_infos = vmwarevif.get_vif_info(self._session, [ 826.135391] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 826.135391] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] for vif in network_info: [ 826.135391] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 826.135391] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] return self._sync_wrapper(fn, *args, **kwargs) [ 826.135391] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 826.135391] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] self.wait() [ 826.135391] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 826.135391] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] self[:] = self._gt.wait() [ 826.135391] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 826.135391] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] return self._exit_event.wait() [ 826.135391] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 826.135987] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] result = hub.switch() [ 826.135987] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 826.135987] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] return self.greenlet.switch() [ 826.135987] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 826.135987] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] result = function(*args, **kwargs) [ 826.135987] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 826.135987] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] return func(*args, **kwargs) [ 826.135987] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 826.135987] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] raise e [ 826.135987] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 826.135987] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] nwinfo = self.network_api.allocate_for_instance( [ 826.135987] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 826.135987] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] created_port_ids = self._update_ports_for_instance( [ 826.136366] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 826.136366] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] with excutils.save_and_reraise_exception(): [ 826.136366] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 826.136366] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] self.force_reraise() [ 826.136366] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 826.136366] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] raise self.value [ 826.136366] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 826.136366] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] updated_port = self._update_port( [ 826.136366] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 826.136366] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] _ensure_no_port_binding_failure(port) [ 826.136366] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 826.136366] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] raise exception.PortBindingFailed(port_id=port['id']) [ 826.136700] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] nova.exception.PortBindingFailed: Binding failed for port f16fa600-42c4-472d-8c4d-468806604d81, please check neutron logs for more information. [ 826.136700] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] [ 826.136700] env[61906]: INFO nova.compute.manager [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Terminating instance [ 826.136700] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Acquiring lock "refresh_cache-34f333be-ea19-499e-ab89-e9f7663ae596" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.136700] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Acquired lock "refresh_cache-34f333be-ea19-499e-ab89-e9f7663ae596" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.136849] env[61906]: DEBUG nova.network.neutron [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 826.165115] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.337664] env[61906]: DEBUG nova.network.neutron [-] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.375365] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.451s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.375892] env[61906]: DEBUG nova.compute.manager [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 826.378476] env[61906]: DEBUG oslo_concurrency.lockutils [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.321s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.379841] env[61906]: INFO nova.compute.claims [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 826.659720] env[61906]: DEBUG nova.network.neutron [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 826.719967] env[61906]: DEBUG nova.network.neutron [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.840722] env[61906]: INFO nova.compute.manager [-] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Took 1.03 seconds to deallocate network for instance. [ 826.843297] env[61906]: DEBUG nova.compute.claims [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 826.843439] env[61906]: DEBUG oslo_concurrency.lockutils [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.891042] env[61906]: DEBUG nova.compute.utils [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 826.892506] env[61906]: DEBUG nova.compute.manager [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 826.892711] env[61906]: DEBUG nova.network.neutron [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 826.943359] env[61906]: DEBUG nova.policy [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '42e109522f074d7fbcc127511917283a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7dc2f7e02ab949f58a1097a67ba9ce06', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 827.205984] env[61906]: DEBUG nova.network.neutron [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Successfully created port: 3644c245-2b19-4580-b700-95677871b770 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 827.222712] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Releasing lock "refresh_cache-34f333be-ea19-499e-ab89-e9f7663ae596" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.223208] env[61906]: DEBUG nova.compute.manager [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 827.223414] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 827.223722] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a9a4190d-5385-4416-8b96-bdba7a9f48cd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.235802] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a5f3d10-7c09-4644-b49b-967d764337f7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.272414] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 34f333be-ea19-499e-ab89-e9f7663ae596 could not be found. [ 827.272414] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 827.272414] env[61906]: INFO nova.compute.manager [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Took 0.05 seconds to destroy the instance on the hypervisor. [ 827.272414] env[61906]: DEBUG oslo.service.loopingcall [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 827.272414] env[61906]: DEBUG nova.compute.manager [-] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 827.272414] env[61906]: DEBUG nova.network.neutron [-] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 827.297696] env[61906]: DEBUG nova.network.neutron [-] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 827.397780] env[61906]: DEBUG nova.compute.manager [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 827.477583] env[61906]: DEBUG nova.compute.manager [req-f56d60bf-63f0-458b-ada2-ea1e4af9a02d req-7f8cc9db-ba83-4217-8839-e6282d8f361c service nova] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Received event network-changed-f16fa600-42c4-472d-8c4d-468806604d81 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 827.477777] env[61906]: DEBUG nova.compute.manager [req-f56d60bf-63f0-458b-ada2-ea1e4af9a02d req-7f8cc9db-ba83-4217-8839-e6282d8f361c service nova] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Refreshing instance network info cache due to event network-changed-f16fa600-42c4-472d-8c4d-468806604d81. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 827.478011] env[61906]: DEBUG oslo_concurrency.lockutils [req-f56d60bf-63f0-458b-ada2-ea1e4af9a02d req-7f8cc9db-ba83-4217-8839-e6282d8f361c service nova] Acquiring lock "refresh_cache-34f333be-ea19-499e-ab89-e9f7663ae596" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.478271] env[61906]: DEBUG oslo_concurrency.lockutils [req-f56d60bf-63f0-458b-ada2-ea1e4af9a02d req-7f8cc9db-ba83-4217-8839-e6282d8f361c service nova] Acquired lock "refresh_cache-34f333be-ea19-499e-ab89-e9f7663ae596" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.478319] env[61906]: DEBUG nova.network.neutron [req-f56d60bf-63f0-458b-ada2-ea1e4af9a02d req-7f8cc9db-ba83-4217-8839-e6282d8f361c service nova] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Refreshing network info cache for port f16fa600-42c4-472d-8c4d-468806604d81 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 827.725174] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9799aa5a-68cd-4b08-b740-99f17c1d45d5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.734096] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a4fce64-cf84-4c6d-9822-3aacb6850a58 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.770751] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75cbcf29-c9a8-4e71-bac6-e5510c329b49 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.779770] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b15ad78-9ae5-47fc-832e-2f3f1b9d0ff4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.799703] env[61906]: DEBUG nova.network.neutron [-] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.801802] env[61906]: DEBUG nova.compute.provider_tree [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 828.013031] env[61906]: DEBUG nova.network.neutron [req-f56d60bf-63f0-458b-ada2-ea1e4af9a02d req-7f8cc9db-ba83-4217-8839-e6282d8f361c service nova] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 828.125158] env[61906]: DEBUG nova.network.neutron [req-f56d60bf-63f0-458b-ada2-ea1e4af9a02d req-7f8cc9db-ba83-4217-8839-e6282d8f361c service nova] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.304675] env[61906]: INFO nova.compute.manager [-] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Took 1.03 seconds to deallocate network for instance. [ 828.305479] env[61906]: DEBUG nova.scheduler.client.report [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 828.312632] env[61906]: DEBUG nova.compute.claims [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 828.312840] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.408889] env[61906]: DEBUG nova.compute.manager [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 828.439517] env[61906]: DEBUG nova.virt.hardware [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 828.441016] env[61906]: DEBUG nova.virt.hardware [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 828.441016] env[61906]: DEBUG nova.virt.hardware [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 828.441016] env[61906]: DEBUG nova.virt.hardware [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 828.441016] env[61906]: DEBUG nova.virt.hardware [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 828.441016] env[61906]: DEBUG nova.virt.hardware [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 828.441195] env[61906]: DEBUG nova.virt.hardware [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 828.441195] env[61906]: DEBUG nova.virt.hardware [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 828.441538] env[61906]: DEBUG nova.virt.hardware [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 828.441893] env[61906]: DEBUG nova.virt.hardware [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 828.442229] env[61906]: DEBUG nova.virt.hardware [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 828.443258] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed913414-ed3b-485d-8620-003416cccaa3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.453319] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9373cb07-2fe9-4282-925b-bfafb7ab472c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.532933] env[61906]: ERROR nova.compute.manager [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3644c245-2b19-4580-b700-95677871b770, please check neutron logs for more information. [ 828.532933] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 828.532933] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 828.532933] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 828.532933] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 828.532933] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 828.532933] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 828.532933] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 828.532933] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 828.532933] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 828.532933] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 828.532933] env[61906]: ERROR nova.compute.manager raise self.value [ 828.532933] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 828.532933] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 828.532933] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 828.532933] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 828.533345] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 828.533345] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 828.533345] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3644c245-2b19-4580-b700-95677871b770, please check neutron logs for more information. [ 828.533345] env[61906]: ERROR nova.compute.manager [ 828.533345] env[61906]: Traceback (most recent call last): [ 828.533345] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 828.533345] env[61906]: listener.cb(fileno) [ 828.533345] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 828.533345] env[61906]: result = function(*args, **kwargs) [ 828.533345] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 828.533345] env[61906]: return func(*args, **kwargs) [ 828.533345] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 828.533345] env[61906]: raise e [ 828.533345] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 828.533345] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 828.533345] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 828.533345] env[61906]: created_port_ids = self._update_ports_for_instance( [ 828.533345] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 828.533345] env[61906]: with excutils.save_and_reraise_exception(): [ 828.533345] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 828.533345] env[61906]: self.force_reraise() [ 828.533345] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 828.533345] env[61906]: raise self.value [ 828.533345] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 828.533345] env[61906]: updated_port = self._update_port( [ 828.533345] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 828.533345] env[61906]: _ensure_no_port_binding_failure(port) [ 828.533345] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 828.533345] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 828.534069] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 3644c245-2b19-4580-b700-95677871b770, please check neutron logs for more information. [ 828.534069] env[61906]: Removing descriptor: 18 [ 828.534069] env[61906]: ERROR nova.compute.manager [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3644c245-2b19-4580-b700-95677871b770, please check neutron logs for more information. [ 828.534069] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Traceback (most recent call last): [ 828.534069] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 828.534069] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] yield resources [ 828.534069] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 828.534069] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] self.driver.spawn(context, instance, image_meta, [ 828.534069] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 828.534069] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 828.534069] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 828.534069] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] vm_ref = self.build_virtual_machine(instance, [ 828.534345] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 828.534345] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] vif_infos = vmwarevif.get_vif_info(self._session, [ 828.534345] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 828.534345] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] for vif in network_info: [ 828.534345] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 828.534345] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] return self._sync_wrapper(fn, *args, **kwargs) [ 828.534345] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 828.534345] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] self.wait() [ 828.534345] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 828.534345] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] self[:] = self._gt.wait() [ 828.534345] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 828.534345] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] return self._exit_event.wait() [ 828.534345] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 828.534638] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] result = hub.switch() [ 828.534638] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 828.534638] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] return self.greenlet.switch() [ 828.534638] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 828.534638] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] result = function(*args, **kwargs) [ 828.534638] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 828.534638] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] return func(*args, **kwargs) [ 828.534638] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 828.534638] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] raise e [ 828.534638] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 828.534638] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] nwinfo = self.network_api.allocate_for_instance( [ 828.534638] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 828.534638] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] created_port_ids = self._update_ports_for_instance( [ 828.534924] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 828.534924] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] with excutils.save_and_reraise_exception(): [ 828.534924] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 828.534924] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] self.force_reraise() [ 828.534924] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 828.534924] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] raise self.value [ 828.534924] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 828.534924] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] updated_port = self._update_port( [ 828.534924] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 828.534924] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] _ensure_no_port_binding_failure(port) [ 828.534924] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 828.534924] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] raise exception.PortBindingFailed(port_id=port['id']) [ 828.535209] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] nova.exception.PortBindingFailed: Binding failed for port 3644c245-2b19-4580-b700-95677871b770, please check neutron logs for more information. [ 828.535209] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] [ 828.535209] env[61906]: INFO nova.compute.manager [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Terminating instance [ 828.536266] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Acquiring lock "refresh_cache-7183e141-9989-4a90-b2fb-2ca0d8e6cda9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.536466] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Acquired lock "refresh_cache-7183e141-9989-4a90-b2fb-2ca0d8e6cda9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.536575] env[61906]: DEBUG nova.network.neutron [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 828.627512] env[61906]: DEBUG oslo_concurrency.lockutils [req-f56d60bf-63f0-458b-ada2-ea1e4af9a02d req-7f8cc9db-ba83-4217-8839-e6282d8f361c service nova] Releasing lock "refresh_cache-34f333be-ea19-499e-ab89-e9f7663ae596" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.627791] env[61906]: DEBUG nova.compute.manager [req-f56d60bf-63f0-458b-ada2-ea1e4af9a02d req-7f8cc9db-ba83-4217-8839-e6282d8f361c service nova] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Received event network-vif-deleted-f16fa600-42c4-472d-8c4d-468806604d81 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 828.814118] env[61906]: DEBUG oslo_concurrency.lockutils [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.435s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.814676] env[61906]: DEBUG nova.compute.manager [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 828.818402] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.757s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.823909] env[61906]: INFO nova.compute.claims [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 829.059778] env[61906]: DEBUG nova.network.neutron [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 829.134993] env[61906]: DEBUG nova.network.neutron [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.320593] env[61906]: DEBUG nova.compute.utils [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 829.324018] env[61906]: DEBUG nova.compute.manager [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 829.324018] env[61906]: DEBUG nova.network.neutron [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 829.367917] env[61906]: DEBUG nova.policy [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e89d45df95134c709b569d06744dea5e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1b33dea0794a48f78b9f519cb269a8c2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 829.518225] env[61906]: DEBUG nova.compute.manager [req-647a0bf5-de8a-4367-a0a7-f4ee79ef262e req-8bfea7fb-3e00-4613-b63b-347ee5b4fe7d service nova] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Received event network-changed-3644c245-2b19-4580-b700-95677871b770 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 829.518476] env[61906]: DEBUG nova.compute.manager [req-647a0bf5-de8a-4367-a0a7-f4ee79ef262e req-8bfea7fb-3e00-4613-b63b-347ee5b4fe7d service nova] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Refreshing instance network info cache due to event network-changed-3644c245-2b19-4580-b700-95677871b770. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 829.518643] env[61906]: DEBUG oslo_concurrency.lockutils [req-647a0bf5-de8a-4367-a0a7-f4ee79ef262e req-8bfea7fb-3e00-4613-b63b-347ee5b4fe7d service nova] Acquiring lock "refresh_cache-7183e141-9989-4a90-b2fb-2ca0d8e6cda9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.637583] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Releasing lock "refresh_cache-7183e141-9989-4a90-b2fb-2ca0d8e6cda9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.638057] env[61906]: DEBUG nova.compute.manager [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 829.638263] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 829.638591] env[61906]: DEBUG oslo_concurrency.lockutils [req-647a0bf5-de8a-4367-a0a7-f4ee79ef262e req-8bfea7fb-3e00-4613-b63b-347ee5b4fe7d service nova] Acquired lock "refresh_cache-7183e141-9989-4a90-b2fb-2ca0d8e6cda9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.638771] env[61906]: DEBUG nova.network.neutron [req-647a0bf5-de8a-4367-a0a7-f4ee79ef262e req-8bfea7fb-3e00-4613-b63b-347ee5b4fe7d service nova] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Refreshing network info cache for port 3644c245-2b19-4580-b700-95677871b770 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 829.640593] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c08679c9-9f1c-4673-845c-d7d8264822e4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.652415] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cb776ce-23e9-447d-beb7-9c0abeb12585 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.678602] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7183e141-9989-4a90-b2fb-2ca0d8e6cda9 could not be found. [ 829.678917] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 829.679076] env[61906]: INFO nova.compute.manager [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 829.679335] env[61906]: DEBUG oslo.service.loopingcall [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.680476] env[61906]: DEBUG nova.compute.manager [-] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 829.680476] env[61906]: DEBUG nova.network.neutron [-] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 829.733385] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "7504dfa6-bedc-4701-b4fc-60e19e742276" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.733606] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "7504dfa6-bedc-4701-b4fc-60e19e742276" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.825096] env[61906]: DEBUG nova.compute.manager [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 829.850151] env[61906]: DEBUG nova.network.neutron [-] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 829.871689] env[61906]: DEBUG nova.network.neutron [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Successfully created port: 7033b729-91d5-4497-adeb-aea522b40942 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 830.147200] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db6f8aca-66c3-442d-a082-f774ab3d9b81 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.155919] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba28a73d-9295-4ce8-89e2-7ea5ea54ee5b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.193125] env[61906]: DEBUG nova.network.neutron [req-647a0bf5-de8a-4367-a0a7-f4ee79ef262e req-8bfea7fb-3e00-4613-b63b-347ee5b4fe7d service nova] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 830.195250] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75a531f0-5e82-4a0d-b1ab-cb187585aab0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.204233] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d9e5538-63c0-41a0-94d7-91ed614a0081 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.218543] env[61906]: DEBUG nova.compute.provider_tree [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 830.333367] env[61906]: DEBUG nova.network.neutron [req-647a0bf5-de8a-4367-a0a7-f4ee79ef262e req-8bfea7fb-3e00-4613-b63b-347ee5b4fe7d service nova] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.359148] env[61906]: DEBUG nova.network.neutron [-] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.725586] env[61906]: DEBUG nova.scheduler.client.report [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 830.833838] env[61906]: DEBUG nova.compute.manager [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 830.836769] env[61906]: DEBUG oslo_concurrency.lockutils [req-647a0bf5-de8a-4367-a0a7-f4ee79ef262e req-8bfea7fb-3e00-4613-b63b-347ee5b4fe7d service nova] Releasing lock "refresh_cache-7183e141-9989-4a90-b2fb-2ca0d8e6cda9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.837137] env[61906]: DEBUG nova.compute.manager [req-647a0bf5-de8a-4367-a0a7-f4ee79ef262e req-8bfea7fb-3e00-4613-b63b-347ee5b4fe7d service nova] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Received event network-vif-deleted-3644c245-2b19-4580-b700-95677871b770 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 830.862407] env[61906]: INFO nova.compute.manager [-] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Took 1.18 seconds to deallocate network for instance. [ 830.866383] env[61906]: DEBUG nova.compute.claims [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 830.866383] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.871792] env[61906]: DEBUG nova.virt.hardware [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 830.873344] env[61906]: DEBUG nova.virt.hardware [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 830.873344] env[61906]: DEBUG nova.virt.hardware [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 830.873344] env[61906]: DEBUG nova.virt.hardware [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 830.873344] env[61906]: DEBUG nova.virt.hardware [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 830.873344] env[61906]: DEBUG nova.virt.hardware [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 830.873573] env[61906]: DEBUG nova.virt.hardware [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 830.873573] env[61906]: DEBUG nova.virt.hardware [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 830.873573] env[61906]: DEBUG nova.virt.hardware [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 830.873573] env[61906]: DEBUG nova.virt.hardware [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 830.873697] env[61906]: DEBUG nova.virt.hardware [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 830.874831] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e0da458-0256-41e6-9a29-faf04941f708 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.883916] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8cfb407-a0e9-4c93-91bb-5b43554d0319 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.997232] env[61906]: ERROR nova.compute.manager [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7033b729-91d5-4497-adeb-aea522b40942, please check neutron logs for more information. [ 830.997232] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 830.997232] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 830.997232] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 830.997232] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 830.997232] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 830.997232] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 830.997232] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 830.997232] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 830.997232] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 830.997232] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 830.997232] env[61906]: ERROR nova.compute.manager raise self.value [ 830.997232] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 830.997232] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 830.997232] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 830.997232] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 830.997703] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 830.997703] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 830.997703] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7033b729-91d5-4497-adeb-aea522b40942, please check neutron logs for more information. [ 830.997703] env[61906]: ERROR nova.compute.manager [ 830.997703] env[61906]: Traceback (most recent call last): [ 830.997703] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 830.997703] env[61906]: listener.cb(fileno) [ 830.997703] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 830.997703] env[61906]: result = function(*args, **kwargs) [ 830.997703] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 830.997703] env[61906]: return func(*args, **kwargs) [ 830.997703] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 830.997703] env[61906]: raise e [ 830.997703] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 830.997703] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 830.997703] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 830.997703] env[61906]: created_port_ids = self._update_ports_for_instance( [ 830.997703] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 830.997703] env[61906]: with excutils.save_and_reraise_exception(): [ 830.997703] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 830.997703] env[61906]: self.force_reraise() [ 830.997703] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 830.997703] env[61906]: raise self.value [ 830.997703] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 830.997703] env[61906]: updated_port = self._update_port( [ 830.997703] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 830.997703] env[61906]: _ensure_no_port_binding_failure(port) [ 830.997703] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 830.997703] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 830.998585] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 7033b729-91d5-4497-adeb-aea522b40942, please check neutron logs for more information. [ 830.998585] env[61906]: Removing descriptor: 18 [ 830.998585] env[61906]: ERROR nova.compute.manager [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7033b729-91d5-4497-adeb-aea522b40942, please check neutron logs for more information. [ 830.998585] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Traceback (most recent call last): [ 830.998585] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 830.998585] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] yield resources [ 830.998585] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 830.998585] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] self.driver.spawn(context, instance, image_meta, [ 830.998585] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 830.998585] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 830.998585] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 830.998585] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] vm_ref = self.build_virtual_machine(instance, [ 830.998921] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 830.998921] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] vif_infos = vmwarevif.get_vif_info(self._session, [ 830.998921] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 830.998921] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] for vif in network_info: [ 830.998921] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 830.998921] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] return self._sync_wrapper(fn, *args, **kwargs) [ 830.998921] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 830.998921] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] self.wait() [ 830.998921] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 830.998921] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] self[:] = self._gt.wait() [ 830.998921] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 830.998921] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] return self._exit_event.wait() [ 830.998921] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 830.999290] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] result = hub.switch() [ 830.999290] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 830.999290] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] return self.greenlet.switch() [ 830.999290] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 830.999290] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] result = function(*args, **kwargs) [ 830.999290] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 830.999290] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] return func(*args, **kwargs) [ 830.999290] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 830.999290] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] raise e [ 830.999290] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 830.999290] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] nwinfo = self.network_api.allocate_for_instance( [ 830.999290] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 830.999290] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] created_port_ids = self._update_ports_for_instance( [ 830.999665] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 830.999665] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] with excutils.save_and_reraise_exception(): [ 830.999665] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 830.999665] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] self.force_reraise() [ 830.999665] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 830.999665] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] raise self.value [ 830.999665] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 830.999665] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] updated_port = self._update_port( [ 830.999665] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 830.999665] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] _ensure_no_port_binding_failure(port) [ 830.999665] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 830.999665] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] raise exception.PortBindingFailed(port_id=port['id']) [ 830.999976] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] nova.exception.PortBindingFailed: Binding failed for port 7033b729-91d5-4497-adeb-aea522b40942, please check neutron logs for more information. [ 830.999976] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] [ 830.999976] env[61906]: INFO nova.compute.manager [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Terminating instance [ 831.000769] env[61906]: DEBUG oslo_concurrency.lockutils [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "refresh_cache-e0afe101-f4a4-4dc8-a666-9a7793dbf8a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.000935] env[61906]: DEBUG oslo_concurrency.lockutils [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquired lock "refresh_cache-e0afe101-f4a4-4dc8-a666-9a7793dbf8a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.001123] env[61906]: DEBUG nova.network.neutron [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 831.233205] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.415s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.233779] env[61906]: DEBUG nova.compute.manager [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 831.236941] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.437s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.238388] env[61906]: INFO nova.compute.claims [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 831.523612] env[61906]: DEBUG nova.network.neutron [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 831.630900] env[61906]: DEBUG nova.network.neutron [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.643704] env[61906]: DEBUG nova.compute.manager [req-75b3c2f2-eb8a-468c-ac00-813819dc3649 req-b420df64-6681-4f5a-8fa9-ea788ded4dc6 service nova] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Received event network-changed-7033b729-91d5-4497-adeb-aea522b40942 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.643704] env[61906]: DEBUG nova.compute.manager [req-75b3c2f2-eb8a-468c-ac00-813819dc3649 req-b420df64-6681-4f5a-8fa9-ea788ded4dc6 service nova] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Refreshing instance network info cache due to event network-changed-7033b729-91d5-4497-adeb-aea522b40942. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 831.643704] env[61906]: DEBUG oslo_concurrency.lockutils [req-75b3c2f2-eb8a-468c-ac00-813819dc3649 req-b420df64-6681-4f5a-8fa9-ea788ded4dc6 service nova] Acquiring lock "refresh_cache-e0afe101-f4a4-4dc8-a666-9a7793dbf8a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.744511] env[61906]: DEBUG nova.compute.utils [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 831.748165] env[61906]: DEBUG nova.compute.manager [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 831.748165] env[61906]: DEBUG nova.network.neutron [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 831.795242] env[61906]: DEBUG nova.policy [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8dfe817014454ba0a23a4998c512739c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ef62024eac3a4333b7c9141ec34c7603', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 832.091297] env[61906]: DEBUG nova.network.neutron [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Successfully created port: 122dafdd-c10c-4440-84d5-755b9239a2ce {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 832.133259] env[61906]: DEBUG oslo_concurrency.lockutils [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Releasing lock "refresh_cache-e0afe101-f4a4-4dc8-a666-9a7793dbf8a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.133702] env[61906]: DEBUG nova.compute.manager [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 832.133900] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 832.134289] env[61906]: DEBUG oslo_concurrency.lockutils [req-75b3c2f2-eb8a-468c-ac00-813819dc3649 req-b420df64-6681-4f5a-8fa9-ea788ded4dc6 service nova] Acquired lock "refresh_cache-e0afe101-f4a4-4dc8-a666-9a7793dbf8a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.134469] env[61906]: DEBUG nova.network.neutron [req-75b3c2f2-eb8a-468c-ac00-813819dc3649 req-b420df64-6681-4f5a-8fa9-ea788ded4dc6 service nova] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Refreshing network info cache for port 7033b729-91d5-4497-adeb-aea522b40942 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 832.135556] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b35738f1-e799-4c5e-b0f3-033327849842 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.146300] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03de940a-5427-4e8a-a552-7dc87dca4073 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.169913] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e0afe101-f4a4-4dc8-a666-9a7793dbf8a6 could not be found. [ 832.170158] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 832.170344] env[61906]: INFO nova.compute.manager [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 832.170582] env[61906]: DEBUG oslo.service.loopingcall [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 832.170805] env[61906]: DEBUG nova.compute.manager [-] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 832.170899] env[61906]: DEBUG nova.network.neutron [-] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 832.190186] env[61906]: DEBUG nova.network.neutron [-] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 832.248763] env[61906]: DEBUG nova.compute.manager [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 832.369074] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Acquiring lock "da493512-d996-4de7-9f47-cadcbc4fbcb9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.369318] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Lock "da493512-d996-4de7-9f47-cadcbc4fbcb9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.555921] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e8c56da-0f2c-4a81-8b14-faafb222da4e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.565075] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f116bd2f-1ac5-4a69-b162-56ee05159106 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.597276] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2b2967-36b2-4ed4-ac78-3b5059fe45a1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.603149] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1300aaac-ead5-4935-8a96-895df21ab39a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.616668] env[61906]: DEBUG nova.compute.provider_tree [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.665394] env[61906]: DEBUG nova.network.neutron [req-75b3c2f2-eb8a-468c-ac00-813819dc3649 req-b420df64-6681-4f5a-8fa9-ea788ded4dc6 service nova] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 832.692818] env[61906]: DEBUG nova.network.neutron [-] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.829028] env[61906]: DEBUG nova.network.neutron [req-75b3c2f2-eb8a-468c-ac00-813819dc3649 req-b420df64-6681-4f5a-8fa9-ea788ded4dc6 service nova] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.120081] env[61906]: DEBUG nova.scheduler.client.report [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 833.164996] env[61906]: ERROR nova.compute.manager [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 122dafdd-c10c-4440-84d5-755b9239a2ce, please check neutron logs for more information. [ 833.164996] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 833.164996] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 833.164996] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 833.164996] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 833.164996] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 833.164996] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 833.164996] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 833.164996] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 833.164996] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 833.164996] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 833.164996] env[61906]: ERROR nova.compute.manager raise self.value [ 833.164996] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 833.164996] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 833.164996] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 833.164996] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 833.165458] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 833.165458] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 833.165458] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 122dafdd-c10c-4440-84d5-755b9239a2ce, please check neutron logs for more information. [ 833.165458] env[61906]: ERROR nova.compute.manager [ 833.165458] env[61906]: Traceback (most recent call last): [ 833.165458] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 833.165458] env[61906]: listener.cb(fileno) [ 833.165458] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 833.165458] env[61906]: result = function(*args, **kwargs) [ 833.165458] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 833.165458] env[61906]: return func(*args, **kwargs) [ 833.165458] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 833.165458] env[61906]: raise e [ 833.165458] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 833.165458] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 833.165458] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 833.165458] env[61906]: created_port_ids = self._update_ports_for_instance( [ 833.165458] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 833.165458] env[61906]: with excutils.save_and_reraise_exception(): [ 833.165458] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 833.165458] env[61906]: self.force_reraise() [ 833.165458] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 833.165458] env[61906]: raise self.value [ 833.165458] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 833.165458] env[61906]: updated_port = self._update_port( [ 833.165458] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 833.165458] env[61906]: _ensure_no_port_binding_failure(port) [ 833.165458] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 833.165458] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 833.166276] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 122dafdd-c10c-4440-84d5-755b9239a2ce, please check neutron logs for more information. [ 833.166276] env[61906]: Removing descriptor: 18 [ 833.195275] env[61906]: INFO nova.compute.manager [-] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Took 1.02 seconds to deallocate network for instance. [ 833.197620] env[61906]: DEBUG nova.compute.claims [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 833.197828] env[61906]: DEBUG oslo_concurrency.lockutils [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.263036] env[61906]: DEBUG nova.compute.manager [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 833.287837] env[61906]: DEBUG nova.virt.hardware [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 833.288094] env[61906]: DEBUG nova.virt.hardware [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 833.288255] env[61906]: DEBUG nova.virt.hardware [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 833.288445] env[61906]: DEBUG nova.virt.hardware [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 833.288633] env[61906]: DEBUG nova.virt.hardware [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 833.288800] env[61906]: DEBUG nova.virt.hardware [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 833.288919] env[61906]: DEBUG nova.virt.hardware [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 833.289089] env[61906]: DEBUG nova.virt.hardware [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 833.289258] env[61906]: DEBUG nova.virt.hardware [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 833.289475] env[61906]: DEBUG nova.virt.hardware [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 833.289593] env[61906]: DEBUG nova.virt.hardware [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 833.290467] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b19e9931-d0ec-4976-85c8-f51320c240e4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.299218] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c560fe9-7349-45c8-8211-abb2fc3ba4ca {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.314183] env[61906]: ERROR nova.compute.manager [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 122dafdd-c10c-4440-84d5-755b9239a2ce, please check neutron logs for more information. [ 833.314183] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] Traceback (most recent call last): [ 833.314183] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 833.314183] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] yield resources [ 833.314183] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 833.314183] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] self.driver.spawn(context, instance, image_meta, [ 833.314183] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 833.314183] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] self._vmops.spawn(context, instance, image_meta, injected_files, [ 833.314183] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 833.314183] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] vm_ref = self.build_virtual_machine(instance, [ 833.314183] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 833.314679] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] vif_infos = vmwarevif.get_vif_info(self._session, [ 833.314679] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 833.314679] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] for vif in network_info: [ 833.314679] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 833.314679] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] return self._sync_wrapper(fn, *args, **kwargs) [ 833.314679] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 833.314679] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] self.wait() [ 833.314679] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 833.314679] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] self[:] = self._gt.wait() [ 833.314679] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 833.314679] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] return self._exit_event.wait() [ 833.314679] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 833.314679] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] current.throw(*self._exc) [ 833.315245] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 833.315245] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] result = function(*args, **kwargs) [ 833.315245] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 833.315245] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] return func(*args, **kwargs) [ 833.315245] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 833.315245] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] raise e [ 833.315245] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 833.315245] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] nwinfo = self.network_api.allocate_for_instance( [ 833.315245] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 833.315245] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] created_port_ids = self._update_ports_for_instance( [ 833.315245] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 833.315245] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] with excutils.save_and_reraise_exception(): [ 833.315245] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 833.315735] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] self.force_reraise() [ 833.315735] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 833.315735] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] raise self.value [ 833.315735] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 833.315735] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] updated_port = self._update_port( [ 833.315735] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 833.315735] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] _ensure_no_port_binding_failure(port) [ 833.315735] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 833.315735] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] raise exception.PortBindingFailed(port_id=port['id']) [ 833.315735] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] nova.exception.PortBindingFailed: Binding failed for port 122dafdd-c10c-4440-84d5-755b9239a2ce, please check neutron logs for more information. [ 833.315735] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] [ 833.315735] env[61906]: INFO nova.compute.manager [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Terminating instance [ 833.316418] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Acquiring lock "refresh_cache-a651bd36-b527-4743-bc6e-5534c629e409" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.316574] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Acquired lock "refresh_cache-a651bd36-b527-4743-bc6e-5534c629e409" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.316760] env[61906]: DEBUG nova.network.neutron [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 833.331559] env[61906]: DEBUG oslo_concurrency.lockutils [req-75b3c2f2-eb8a-468c-ac00-813819dc3649 req-b420df64-6681-4f5a-8fa9-ea788ded4dc6 service nova] Releasing lock "refresh_cache-e0afe101-f4a4-4dc8-a666-9a7793dbf8a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.331809] env[61906]: DEBUG nova.compute.manager [req-75b3c2f2-eb8a-468c-ac00-813819dc3649 req-b420df64-6681-4f5a-8fa9-ea788ded4dc6 service nova] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Received event network-vif-deleted-7033b729-91d5-4497-adeb-aea522b40942 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.626767] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.390s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.627419] env[61906]: DEBUG nova.compute.manager [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 833.630695] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.956s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.667277] env[61906]: DEBUG nova.compute.manager [req-7af6f578-ac38-4e7f-af1e-dd55ee6e5f03 req-250da652-c878-481e-a1c6-7dd48a9a2ccc service nova] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Received event network-changed-122dafdd-c10c-4440-84d5-755b9239a2ce {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.667376] env[61906]: DEBUG nova.compute.manager [req-7af6f578-ac38-4e7f-af1e-dd55ee6e5f03 req-250da652-c878-481e-a1c6-7dd48a9a2ccc service nova] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Refreshing instance network info cache due to event network-changed-122dafdd-c10c-4440-84d5-755b9239a2ce. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 833.667560] env[61906]: DEBUG oslo_concurrency.lockutils [req-7af6f578-ac38-4e7f-af1e-dd55ee6e5f03 req-250da652-c878-481e-a1c6-7dd48a9a2ccc service nova] Acquiring lock "refresh_cache-a651bd36-b527-4743-bc6e-5534c629e409" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.837440] env[61906]: DEBUG nova.network.neutron [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 833.948285] env[61906]: DEBUG nova.network.neutron [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.135422] env[61906]: DEBUG nova.compute.utils [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 834.139756] env[61906]: DEBUG nova.compute.manager [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 834.139756] env[61906]: DEBUG nova.network.neutron [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 834.193202] env[61906]: DEBUG nova.policy [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '172e2bcc683b47fa8750be3f50c068b1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '30b3ffdafa384a4b933c53af3fe6ffc1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 834.386765] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4c1f217-51a6-44ec-b985-5226de29e0db {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.396277] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e980c061-5e99-46ff-9566-2e56f4b94e4c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.427489] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fec557d-4f0e-4680-a1a0-fa9e2ae89da0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.435674] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-149bf4fd-8853-4156-b544-3e6df784ae5b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.451286] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Releasing lock "refresh_cache-a651bd36-b527-4743-bc6e-5534c629e409" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.451286] env[61906]: DEBUG nova.compute.manager [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 834.452369] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 834.452369] env[61906]: DEBUG nova.compute.provider_tree [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 834.453230] env[61906]: DEBUG oslo_concurrency.lockutils [req-7af6f578-ac38-4e7f-af1e-dd55ee6e5f03 req-250da652-c878-481e-a1c6-7dd48a9a2ccc service nova] Acquired lock "refresh_cache-a651bd36-b527-4743-bc6e-5534c629e409" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.457158] env[61906]: DEBUG nova.network.neutron [req-7af6f578-ac38-4e7f-af1e-dd55ee6e5f03 req-250da652-c878-481e-a1c6-7dd48a9a2ccc service nova] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Refreshing network info cache for port 122dafdd-c10c-4440-84d5-755b9239a2ce {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 834.458118] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-68895ce3-00ef-4a18-b28e-b6b70a3b2f65 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.471101] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95c9f226-a6a1-4b1e-967d-b90f8dd1e3b4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.485092] env[61906]: DEBUG nova.network.neutron [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Successfully created port: 21804744-845d-412e-b7c7-84b67dbbb153 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 834.501872] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a651bd36-b527-4743-bc6e-5534c629e409 could not be found. [ 834.502146] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 834.502333] env[61906]: INFO nova.compute.manager [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Took 0.05 seconds to destroy the instance on the hypervisor. [ 834.502647] env[61906]: DEBUG oslo.service.loopingcall [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.502890] env[61906]: DEBUG nova.compute.manager [-] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 834.502986] env[61906]: DEBUG nova.network.neutron [-] [instance: a651bd36-b527-4743-bc6e-5534c629e409] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 834.519315] env[61906]: DEBUG nova.network.neutron [-] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 834.645429] env[61906]: DEBUG nova.compute.manager [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 834.963314] env[61906]: DEBUG nova.scheduler.client.report [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 834.983389] env[61906]: DEBUG nova.network.neutron [req-7af6f578-ac38-4e7f-af1e-dd55ee6e5f03 req-250da652-c878-481e-a1c6-7dd48a9a2ccc service nova] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 835.022035] env[61906]: DEBUG nova.network.neutron [-] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.121702] env[61906]: DEBUG nova.network.neutron [req-7af6f578-ac38-4e7f-af1e-dd55ee6e5f03 req-250da652-c878-481e-a1c6-7dd48a9a2ccc service nova] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.468533] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.838s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.469250] env[61906]: ERROR nova.compute.manager [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 04cfd791-c3c1-4e94-8768-90499c8e9ca0, please check neutron logs for more information. [ 835.469250] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Traceback (most recent call last): [ 835.469250] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 835.469250] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] self.driver.spawn(context, instance, image_meta, [ 835.469250] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 835.469250] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 835.469250] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 835.469250] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] vm_ref = self.build_virtual_machine(instance, [ 835.469250] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 835.469250] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] vif_infos = vmwarevif.get_vif_info(self._session, [ 835.469250] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 835.469550] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] for vif in network_info: [ 835.469550] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 835.469550] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] return self._sync_wrapper(fn, *args, **kwargs) [ 835.469550] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 835.469550] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] self.wait() [ 835.469550] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 835.469550] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] self[:] = self._gt.wait() [ 835.469550] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 835.469550] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] return self._exit_event.wait() [ 835.469550] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 835.469550] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] result = hub.switch() [ 835.469550] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 835.469550] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] return self.greenlet.switch() [ 835.469837] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 835.469837] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] result = function(*args, **kwargs) [ 835.469837] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 835.469837] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] return func(*args, **kwargs) [ 835.469837] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 835.469837] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] raise e [ 835.469837] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 835.469837] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] nwinfo = self.network_api.allocate_for_instance( [ 835.469837] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 835.469837] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] created_port_ids = self._update_ports_for_instance( [ 835.469837] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 835.469837] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] with excutils.save_and_reraise_exception(): [ 835.469837] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 835.470152] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] self.force_reraise() [ 835.470152] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 835.470152] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] raise self.value [ 835.470152] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 835.470152] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] updated_port = self._update_port( [ 835.470152] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 835.470152] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] _ensure_no_port_binding_failure(port) [ 835.470152] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 835.470152] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] raise exception.PortBindingFailed(port_id=port['id']) [ 835.470152] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] nova.exception.PortBindingFailed: Binding failed for port 04cfd791-c3c1-4e94-8768-90499c8e9ca0, please check neutron logs for more information. [ 835.470152] env[61906]: ERROR nova.compute.manager [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] [ 835.473576] env[61906]: DEBUG nova.compute.utils [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Binding failed for port 04cfd791-c3c1-4e94-8768-90499c8e9ca0, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 835.474897] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.323s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.478973] env[61906]: DEBUG nova.compute.manager [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Build of instance b935c7ba-7716-443b-b5da-9ee03cef793a was re-scheduled: Binding failed for port 04cfd791-c3c1-4e94-8768-90499c8e9ca0, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 835.479530] env[61906]: DEBUG nova.compute.manager [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 835.479780] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Acquiring lock "refresh_cache-b935c7ba-7716-443b-b5da-9ee03cef793a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.479934] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Acquired lock "refresh_cache-b935c7ba-7716-443b-b5da-9ee03cef793a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.480112] env[61906]: DEBUG nova.network.neutron [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 835.524951] env[61906]: INFO nova.compute.manager [-] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Took 1.02 seconds to deallocate network for instance. [ 835.527050] env[61906]: DEBUG nova.compute.claims [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 835.527236] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.535959] env[61906]: ERROR nova.compute.manager [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 21804744-845d-412e-b7c7-84b67dbbb153, please check neutron logs for more information. [ 835.535959] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 835.535959] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 835.535959] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 835.535959] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 835.535959] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 835.535959] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 835.535959] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 835.535959] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 835.535959] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 835.535959] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 835.535959] env[61906]: ERROR nova.compute.manager raise self.value [ 835.535959] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 835.535959] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 835.535959] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 835.535959] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 835.536395] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 835.536395] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 835.536395] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 21804744-845d-412e-b7c7-84b67dbbb153, please check neutron logs for more information. [ 835.536395] env[61906]: ERROR nova.compute.manager [ 835.536395] env[61906]: Traceback (most recent call last): [ 835.536395] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 835.536395] env[61906]: listener.cb(fileno) [ 835.536395] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 835.536395] env[61906]: result = function(*args, **kwargs) [ 835.536395] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 835.536395] env[61906]: return func(*args, **kwargs) [ 835.536395] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 835.536395] env[61906]: raise e [ 835.536395] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 835.536395] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 835.536395] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 835.536395] env[61906]: created_port_ids = self._update_ports_for_instance( [ 835.536395] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 835.536395] env[61906]: with excutils.save_and_reraise_exception(): [ 835.536395] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 835.536395] env[61906]: self.force_reraise() [ 835.536395] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 835.536395] env[61906]: raise self.value [ 835.536395] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 835.536395] env[61906]: updated_port = self._update_port( [ 835.536395] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 835.536395] env[61906]: _ensure_no_port_binding_failure(port) [ 835.536395] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 835.536395] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 835.537120] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 21804744-845d-412e-b7c7-84b67dbbb153, please check neutron logs for more information. [ 835.537120] env[61906]: Removing descriptor: 18 [ 835.624549] env[61906]: DEBUG oslo_concurrency.lockutils [req-7af6f578-ac38-4e7f-af1e-dd55ee6e5f03 req-250da652-c878-481e-a1c6-7dd48a9a2ccc service nova] Releasing lock "refresh_cache-a651bd36-b527-4743-bc6e-5534c629e409" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.624885] env[61906]: DEBUG nova.compute.manager [req-7af6f578-ac38-4e7f-af1e-dd55ee6e5f03 req-250da652-c878-481e-a1c6-7dd48a9a2ccc service nova] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Received event network-vif-deleted-122dafdd-c10c-4440-84d5-755b9239a2ce {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 835.652858] env[61906]: DEBUG nova.compute.manager [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 835.677422] env[61906]: DEBUG nova.virt.hardware [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 835.677723] env[61906]: DEBUG nova.virt.hardware [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 835.677861] env[61906]: DEBUG nova.virt.hardware [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 835.678063] env[61906]: DEBUG nova.virt.hardware [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 835.678221] env[61906]: DEBUG nova.virt.hardware [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 835.678369] env[61906]: DEBUG nova.virt.hardware [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 835.678575] env[61906]: DEBUG nova.virt.hardware [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 835.678735] env[61906]: DEBUG nova.virt.hardware [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 835.678902] env[61906]: DEBUG nova.virt.hardware [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 835.679079] env[61906]: DEBUG nova.virt.hardware [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 835.679260] env[61906]: DEBUG nova.virt.hardware [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 835.680171] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-428bab9c-ed7e-4eb5-beb7-390142ed9e44 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.688656] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0bfc345-2afb-4c05-ab72-ac279c12152a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.704390] env[61906]: ERROR nova.compute.manager [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 21804744-845d-412e-b7c7-84b67dbbb153, please check neutron logs for more information. [ 835.704390] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Traceback (most recent call last): [ 835.704390] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 835.704390] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] yield resources [ 835.704390] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 835.704390] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] self.driver.spawn(context, instance, image_meta, [ 835.704390] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 835.704390] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 835.704390] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 835.704390] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] vm_ref = self.build_virtual_machine(instance, [ 835.704390] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 835.704739] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] vif_infos = vmwarevif.get_vif_info(self._session, [ 835.704739] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 835.704739] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] for vif in network_info: [ 835.704739] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 835.704739] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] return self._sync_wrapper(fn, *args, **kwargs) [ 835.704739] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 835.704739] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] self.wait() [ 835.704739] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 835.704739] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] self[:] = self._gt.wait() [ 835.704739] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 835.704739] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] return self._exit_event.wait() [ 835.704739] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 835.704739] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] current.throw(*self._exc) [ 835.705096] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 835.705096] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] result = function(*args, **kwargs) [ 835.705096] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 835.705096] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] return func(*args, **kwargs) [ 835.705096] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 835.705096] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] raise e [ 835.705096] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 835.705096] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] nwinfo = self.network_api.allocate_for_instance( [ 835.705096] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 835.705096] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] created_port_ids = self._update_ports_for_instance( [ 835.705096] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 835.705096] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] with excutils.save_and_reraise_exception(): [ 835.705096] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 835.705508] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] self.force_reraise() [ 835.705508] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 835.705508] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] raise self.value [ 835.705508] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 835.705508] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] updated_port = self._update_port( [ 835.705508] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 835.705508] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] _ensure_no_port_binding_failure(port) [ 835.705508] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 835.705508] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] raise exception.PortBindingFailed(port_id=port['id']) [ 835.705508] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] nova.exception.PortBindingFailed: Binding failed for port 21804744-845d-412e-b7c7-84b67dbbb153, please check neutron logs for more information. [ 835.705508] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] [ 835.705508] env[61906]: INFO nova.compute.manager [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Terminating instance [ 835.707592] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Acquiring lock "refresh_cache-976021e6-0e38-4295-bdd4-83f00a7b7be0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.707792] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Acquired lock "refresh_cache-976021e6-0e38-4295-bdd4-83f00a7b7be0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.707970] env[61906]: DEBUG nova.network.neutron [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 835.715515] env[61906]: DEBUG nova.compute.manager [req-036bf092-64ad-4c7b-88f0-b5f82d92ac14 req-e17717e8-72be-42ab-82ce-d64accdaf8d2 service nova] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Received event network-changed-21804744-845d-412e-b7c7-84b67dbbb153 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 835.715707] env[61906]: DEBUG nova.compute.manager [req-036bf092-64ad-4c7b-88f0-b5f82d92ac14 req-e17717e8-72be-42ab-82ce-d64accdaf8d2 service nova] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Refreshing instance network info cache due to event network-changed-21804744-845d-412e-b7c7-84b67dbbb153. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 835.716092] env[61906]: DEBUG oslo_concurrency.lockutils [req-036bf092-64ad-4c7b-88f0-b5f82d92ac14 req-e17717e8-72be-42ab-82ce-d64accdaf8d2 service nova] Acquiring lock "refresh_cache-976021e6-0e38-4295-bdd4-83f00a7b7be0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.010519] env[61906]: DEBUG nova.network.neutron [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 836.092304] env[61906]: DEBUG nova.network.neutron [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.236063] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-626ee404-5077-46be-9a10-b12e1ff83334 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.239026] env[61906]: DEBUG nova.network.neutron [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 836.247222] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e869fd1-bde5-491d-8dda-53d33d65a3b0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.280703] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6fefac6-ba74-47e2-a09c-7af0d31fd218 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.289214] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79be8e0f-a294-4dbb-935e-0aa8b8b97b50 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.303260] env[61906]: DEBUG nova.compute.provider_tree [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 836.325323] env[61906]: DEBUG nova.network.neutron [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.597153] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Releasing lock "refresh_cache-b935c7ba-7716-443b-b5da-9ee03cef793a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.597499] env[61906]: DEBUG nova.compute.manager [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 836.597586] env[61906]: DEBUG nova.compute.manager [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 836.597753] env[61906]: DEBUG nova.network.neutron [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 836.613102] env[61906]: DEBUG nova.network.neutron [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 836.806434] env[61906]: DEBUG nova.scheduler.client.report [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 836.827801] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Releasing lock "refresh_cache-976021e6-0e38-4295-bdd4-83f00a7b7be0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.828478] env[61906]: DEBUG nova.compute.manager [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 836.828705] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 836.829048] env[61906]: DEBUG oslo_concurrency.lockutils [req-036bf092-64ad-4c7b-88f0-b5f82d92ac14 req-e17717e8-72be-42ab-82ce-d64accdaf8d2 service nova] Acquired lock "refresh_cache-976021e6-0e38-4295-bdd4-83f00a7b7be0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.829235] env[61906]: DEBUG nova.network.neutron [req-036bf092-64ad-4c7b-88f0-b5f82d92ac14 req-e17717e8-72be-42ab-82ce-d64accdaf8d2 service nova] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Refreshing network info cache for port 21804744-845d-412e-b7c7-84b67dbbb153 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 836.830413] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c691c9aa-d85a-4ae5-9900-f3a08580e820 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.840958] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c51d99f1-9697-4017-96ec-6779efc296fb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.864398] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 976021e6-0e38-4295-bdd4-83f00a7b7be0 could not be found. [ 836.864622] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 836.864802] env[61906]: INFO nova.compute.manager [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 836.865205] env[61906]: DEBUG oslo.service.loopingcall [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 836.865514] env[61906]: DEBUG nova.compute.manager [-] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 836.865613] env[61906]: DEBUG nova.network.neutron [-] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 836.880405] env[61906]: DEBUG nova.network.neutron [-] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 837.116860] env[61906]: DEBUG nova.network.neutron [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.311160] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.836s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.311849] env[61906]: ERROR nova.compute.manager [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 88d10958-f4d5-4378-98b5-dae42820a117, please check neutron logs for more information. [ 837.311849] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Traceback (most recent call last): [ 837.311849] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 837.311849] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] self.driver.spawn(context, instance, image_meta, [ 837.311849] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 837.311849] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 837.311849] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 837.311849] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] vm_ref = self.build_virtual_machine(instance, [ 837.311849] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 837.311849] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] vif_infos = vmwarevif.get_vif_info(self._session, [ 837.311849] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 837.312158] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] for vif in network_info: [ 837.312158] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 837.312158] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] return self._sync_wrapper(fn, *args, **kwargs) [ 837.312158] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 837.312158] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] self.wait() [ 837.312158] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 837.312158] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] self[:] = self._gt.wait() [ 837.312158] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 837.312158] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] return self._exit_event.wait() [ 837.312158] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 837.312158] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] current.throw(*self._exc) [ 837.312158] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 837.312158] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] result = function(*args, **kwargs) [ 837.312520] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 837.312520] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] return func(*args, **kwargs) [ 837.312520] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 837.312520] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] raise e [ 837.312520] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 837.312520] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] nwinfo = self.network_api.allocate_for_instance( [ 837.312520] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 837.312520] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] created_port_ids = self._update_ports_for_instance( [ 837.312520] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 837.312520] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] with excutils.save_and_reraise_exception(): [ 837.312520] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 837.312520] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] self.force_reraise() [ 837.312520] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 837.312926] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] raise self.value [ 837.312926] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 837.312926] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] updated_port = self._update_port( [ 837.312926] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 837.312926] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] _ensure_no_port_binding_failure(port) [ 837.312926] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 837.312926] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] raise exception.PortBindingFailed(port_id=port['id']) [ 837.312926] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] nova.exception.PortBindingFailed: Binding failed for port 88d10958-f4d5-4378-98b5-dae42820a117, please check neutron logs for more information. [ 837.312926] env[61906]: ERROR nova.compute.manager [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] [ 837.312926] env[61906]: DEBUG nova.compute.utils [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Binding failed for port 88d10958-f4d5-4378-98b5-dae42820a117, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 837.313898] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 18.095s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.315976] env[61906]: DEBUG nova.compute.manager [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Build of instance 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3 was re-scheduled: Binding failed for port 88d10958-f4d5-4378-98b5-dae42820a117, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 837.316405] env[61906]: DEBUG nova.compute.manager [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 837.316670] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Acquiring lock "refresh_cache-3923d04e-5b50-4cfc-b3da-4b4eada4c8e3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.316870] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Acquired lock "refresh_cache-3923d04e-5b50-4cfc-b3da-4b4eada4c8e3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.316986] env[61906]: DEBUG nova.network.neutron [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 837.356270] env[61906]: DEBUG nova.network.neutron [req-036bf092-64ad-4c7b-88f0-b5f82d92ac14 req-e17717e8-72be-42ab-82ce-d64accdaf8d2 service nova] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 837.383044] env[61906]: DEBUG nova.network.neutron [-] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.439353] env[61906]: DEBUG nova.network.neutron [req-036bf092-64ad-4c7b-88f0-b5f82d92ac14 req-e17717e8-72be-42ab-82ce-d64accdaf8d2 service nova] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.618259] env[61906]: INFO nova.compute.manager [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] [instance: b935c7ba-7716-443b-b5da-9ee03cef793a] Took 1.02 seconds to deallocate network for instance. [ 837.840023] env[61906]: DEBUG nova.network.neutron [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 837.888021] env[61906]: INFO nova.compute.manager [-] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Took 1.02 seconds to deallocate network for instance. [ 837.888021] env[61906]: DEBUG nova.compute.claims [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 837.888021] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.932140] env[61906]: DEBUG nova.network.neutron [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.941854] env[61906]: DEBUG oslo_concurrency.lockutils [req-036bf092-64ad-4c7b-88f0-b5f82d92ac14 req-e17717e8-72be-42ab-82ce-d64accdaf8d2 service nova] Releasing lock "refresh_cache-976021e6-0e38-4295-bdd4-83f00a7b7be0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.942134] env[61906]: DEBUG nova.compute.manager [req-036bf092-64ad-4c7b-88f0-b5f82d92ac14 req-e17717e8-72be-42ab-82ce-d64accdaf8d2 service nova] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Received event network-vif-deleted-21804744-845d-412e-b7c7-84b67dbbb153 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 838.346100] env[61906]: WARNING nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 7e452659-0a5e-4740-b7ed-1fd7b1917dac is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 838.438827] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Releasing lock "refresh_cache-3923d04e-5b50-4cfc-b3da-4b4eada4c8e3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.439189] env[61906]: DEBUG nova.compute.manager [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 838.439292] env[61906]: DEBUG nova.compute.manager [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 838.439464] env[61906]: DEBUG nova.network.neutron [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 838.456498] env[61906]: DEBUG nova.network.neutron [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 838.649862] env[61906]: INFO nova.scheduler.client.report [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Deleted allocations for instance b935c7ba-7716-443b-b5da-9ee03cef793a [ 838.855037] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance b935c7ba-7716-443b-b5da-9ee03cef793a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 838.959849] env[61906]: DEBUG nova.network.neutron [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.158851] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b6a90903-7ef4-4deb-90a6-0023e734d5f3 tempest-ServerRescueTestJSON-1226082340 tempest-ServerRescueTestJSON-1226082340-project-member] Lock "b935c7ba-7716-443b-b5da-9ee03cef793a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 185.315s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.358096] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 839.358324] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance df9c9d62-8dc5-452c-85ed-846dcbd76014 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 839.358405] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 52a387d2-4090-4eba-9a77-98c37f152b75 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 839.358530] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 34f333be-ea19-499e-ab89-e9f7663ae596 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 839.358681] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 7183e141-9989-4a90-b2fb-2ca0d8e6cda9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 839.358818] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance e0afe101-f4a4-4dc8-a666-9a7793dbf8a6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 839.358950] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance a651bd36-b527-4743-bc6e-5534c629e409 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 839.359042] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 976021e6-0e38-4295-bdd4-83f00a7b7be0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 839.462453] env[61906]: INFO nova.compute.manager [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] [instance: 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3] Took 1.02 seconds to deallocate network for instance. [ 839.661563] env[61906]: DEBUG nova.compute.manager [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 839.865636] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 7e05fee8-842f-4e03-853a-58bac5a9ac0e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 840.185946] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.371524] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 9f1253e6-4727-4dca-bad2-04f0c5424cdb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 840.504323] env[61906]: INFO nova.scheduler.client.report [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Deleted allocations for instance 3923d04e-5b50-4cfc-b3da-4b4eada4c8e3 [ 840.875157] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 80a6a0b8-17c3-46e8-8861-6d1759a19111 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 841.017673] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0d9d5b2-3693-444b-9007-b87d761fd1ef tempest-ServerActionsTestOtherB-1127884011 tempest-ServerActionsTestOtherB-1127884011-project-member] Lock "3923d04e-5b50-4cfc-b3da-4b4eada4c8e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 176.505s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.379627] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance d60575a6-af13-4d81-95d1-081d0f3e2ce6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 841.520565] env[61906]: DEBUG nova.compute.manager [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 841.882554] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 842.042688] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.386262] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 842.894523] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 75b9738e-e9b6-435a-88bb-851982828d36 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 843.392860] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance cb34a0ce-d274-4167-8367-e19809cc9e4d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 843.896019] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance cc8870d0-eafb-4f73-aa79-b98f51370237 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 844.399285] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 7504dfa6-bedc-4701-b4fc-60e19e742276 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 844.902239] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance da493512-d996-4de7-9f47-cadcbc4fbcb9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 844.902478] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 844.902659] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 845.104160] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f8456eb-9a05-4d81-b33e-1a9a39c59291 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.112272] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e326f4e4-1a05-4bab-bfea-205bda93bc9a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.141453] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d31efb46-368c-4d61-a62c-73e005f5029b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.149028] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f7f0ee5-a582-4c20-bec5-c2d3f37958aa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.162717] env[61906]: DEBUG nova.compute.provider_tree [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 845.665999] env[61906]: DEBUG nova.scheduler.client.report [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 846.172847] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 846.173224] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 8.859s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.173434] env[61906]: DEBUG oslo_concurrency.lockutils [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 25.014s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.173635] env[61906]: DEBUG nova.objects.instance [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61906) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 847.155809] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 847.155987] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Starting heal instance info cache {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 847.156134] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Rebuilding the list of instances to heal {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 847.182374] env[61906]: DEBUG oslo_concurrency.lockutils [None req-861d465c-5bb3-456e-be1f-361e8ba11110 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.183750] env[61906]: DEBUG oslo_concurrency.lockutils [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.655s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.660419] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 847.660603] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 847.660791] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 847.660960] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 847.661132] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 847.661289] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 847.661440] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 847.677256] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "refresh_cache-7e452659-0a5e-4740-b7ed-1fd7b1917dac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.677423] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquired lock "refresh_cache-7e452659-0a5e-4740-b7ed-1fd7b1917dac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.677591] env[61906]: DEBUG nova.network.neutron [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Forcefully refreshing network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 847.677768] env[61906]: DEBUG nova.objects.instance [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lazy-loading 'info_cache' on Instance uuid 7e452659-0a5e-4740-b7ed-1fd7b1917dac {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 847.906641] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7d9bf9b-1d07-4672-abe9-0a6ddb49eea4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.914759] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5cdf65c-a7b3-45f8-8a8b-5821b6b1a31b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.945624] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9618846a-ec00-4888-8614-c666012a9a57 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.952832] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03023973-fa71-41fa-8c10-ed0562547553 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.965588] env[61906]: DEBUG nova.compute.provider_tree [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.469125] env[61906]: DEBUG nova.scheduler.client.report [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 848.699126] env[61906]: DEBUG nova.network.neutron [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 848.974513] env[61906]: DEBUG oslo_concurrency.lockutils [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.791s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.975192] env[61906]: ERROR nova.compute.manager [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5a9efc88-6a93-4307-859b-8286ddc55f73, please check neutron logs for more information. [ 848.975192] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Traceback (most recent call last): [ 848.975192] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 848.975192] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] self.driver.spawn(context, instance, image_meta, [ 848.975192] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 848.975192] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] self._vmops.spawn(context, instance, image_meta, injected_files, [ 848.975192] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 848.975192] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] vm_ref = self.build_virtual_machine(instance, [ 848.975192] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 848.975192] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] vif_infos = vmwarevif.get_vif_info(self._session, [ 848.975192] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 848.975535] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] for vif in network_info: [ 848.975535] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 848.975535] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] return self._sync_wrapper(fn, *args, **kwargs) [ 848.975535] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 848.975535] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] self.wait() [ 848.975535] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 848.975535] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] self[:] = self._gt.wait() [ 848.975535] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 848.975535] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] return self._exit_event.wait() [ 848.975535] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 848.975535] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] current.throw(*self._exc) [ 848.975535] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 848.975535] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] result = function(*args, **kwargs) [ 848.975898] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 848.975898] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] return func(*args, **kwargs) [ 848.975898] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 848.975898] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] raise e [ 848.975898] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 848.975898] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] nwinfo = self.network_api.allocate_for_instance( [ 848.975898] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 848.975898] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] created_port_ids = self._update_ports_for_instance( [ 848.975898] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 848.975898] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] with excutils.save_and_reraise_exception(): [ 848.975898] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 848.975898] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] self.force_reraise() [ 848.975898] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 848.976278] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] raise self.value [ 848.976278] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 848.976278] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] updated_port = self._update_port( [ 848.976278] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 848.976278] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] _ensure_no_port_binding_failure(port) [ 848.976278] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 848.976278] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] raise exception.PortBindingFailed(port_id=port['id']) [ 848.976278] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] nova.exception.PortBindingFailed: Binding failed for port 5a9efc88-6a93-4307-859b-8286ddc55f73, please check neutron logs for more information. [ 848.976278] env[61906]: ERROR nova.compute.manager [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] [ 848.976278] env[61906]: DEBUG nova.compute.utils [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Binding failed for port 5a9efc88-6a93-4307-859b-8286ddc55f73, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 848.977167] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.061s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.978673] env[61906]: INFO nova.compute.claims [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 848.981259] env[61906]: DEBUG nova.compute.manager [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Build of instance df9c9d62-8dc5-452c-85ed-846dcbd76014 was re-scheduled: Binding failed for port 5a9efc88-6a93-4307-859b-8286ddc55f73, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 848.981674] env[61906]: DEBUG nova.compute.manager [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 848.981926] env[61906]: DEBUG oslo_concurrency.lockutils [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Acquiring lock "refresh_cache-df9c9d62-8dc5-452c-85ed-846dcbd76014" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.982088] env[61906]: DEBUG oslo_concurrency.lockutils [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Acquired lock "refresh_cache-df9c9d62-8dc5-452c-85ed-846dcbd76014" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.982251] env[61906]: DEBUG nova.network.neutron [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 849.263584] env[61906]: DEBUG nova.network.neutron [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.501957] env[61906]: DEBUG nova.network.neutron [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 849.583951] env[61906]: DEBUG nova.network.neutron [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.765860] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Releasing lock "refresh_cache-7e452659-0a5e-4740-b7ed-1fd7b1917dac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.766115] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Updated the network info_cache for instance {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 849.766324] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 850.087268] env[61906]: DEBUG oslo_concurrency.lockutils [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Releasing lock "refresh_cache-df9c9d62-8dc5-452c-85ed-846dcbd76014" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.087386] env[61906]: DEBUG nova.compute.manager [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 850.087647] env[61906]: DEBUG nova.compute.manager [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 850.087844] env[61906]: DEBUG nova.network.neutron [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 850.105487] env[61906]: DEBUG nova.network.neutron [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 850.203636] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8299b219-d3a4-4cc4-8479-65aa25a4d99b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.212252] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-affc8a48-18b1-49b5-a2f9-1de60aa4a9d4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.242626] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c34fe8b2-1f9e-4c1e-b2ba-1dd5f02df157 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.250796] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37c870e0-3d5d-4450-a512-3e04c00e8ba9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.265260] env[61906]: DEBUG nova.compute.provider_tree [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.608580] env[61906]: DEBUG nova.network.neutron [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.768037] env[61906]: DEBUG nova.scheduler.client.report [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 851.111765] env[61906]: INFO nova.compute.manager [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] [instance: df9c9d62-8dc5-452c-85ed-846dcbd76014] Took 1.02 seconds to deallocate network for instance. [ 851.273086] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.296s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.273653] env[61906]: DEBUG nova.compute.manager [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 851.276268] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.111s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.276481] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.278468] env[61906]: DEBUG oslo_concurrency.lockutils [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.435s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.301972] env[61906]: INFO nova.scheduler.client.report [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Deleted allocations for instance 7e452659-0a5e-4740-b7ed-1fd7b1917dac [ 851.783517] env[61906]: DEBUG nova.compute.utils [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 851.789178] env[61906]: DEBUG nova.compute.manager [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 851.789375] env[61906]: DEBUG nova.network.neutron [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 851.810420] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c4efa47-e2db-4899-bf17-d804ad2d5fb8 tempest-ServerShowV257Test-1566975951 tempest-ServerShowV257Test-1566975951-project-member] Lock "7e452659-0a5e-4740-b7ed-1fd7b1917dac" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.466s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.871247] env[61906]: DEBUG nova.policy [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e883fb94e9b74e54925677b3f9466e26', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c15bc344a5440d491b47b9fdedb6869', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 852.008009] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81cbf3b9-24ca-4c42-ba21-692e40929017 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.016748] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-567e5218-aac9-4967-871e-b511216d230b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.059623] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41626cb8-693f-48a1-9ce5-c8e0c397adc2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.068964] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75fa1495-2331-4d14-bcc8-a3bcc8d5eb0d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.083839] env[61906]: DEBUG nova.compute.provider_tree [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 852.161391] env[61906]: INFO nova.scheduler.client.report [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Deleted allocations for instance df9c9d62-8dc5-452c-85ed-846dcbd76014 [ 852.188515] env[61906]: DEBUG nova.network.neutron [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Successfully created port: d5ae18fb-369d-4c5b-9efe-a7d2898521e9 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 852.290748] env[61906]: DEBUG nova.compute.manager [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 852.586748] env[61906]: DEBUG nova.network.neutron [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Successfully created port: c156b7ea-028a-4b34-93cd-25d384ab5e76 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 852.589125] env[61906]: DEBUG nova.scheduler.client.report [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 852.675281] env[61906]: DEBUG oslo_concurrency.lockutils [None req-555d17b9-d5e8-40a8-88d1-7729b7eb862d tempest-ServerActionsTestOtherA-1069435418 tempest-ServerActionsTestOtherA-1069435418-project-member] Lock "df9c9d62-8dc5-452c-85ed-846dcbd76014" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 165.285s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.094701] env[61906]: DEBUG oslo_concurrency.lockutils [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.816s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.095400] env[61906]: ERROR nova.compute.manager [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 018c3cd4-50ad-4134-8ee4-c8aff36ab617, please check neutron logs for more information. [ 853.095400] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Traceback (most recent call last): [ 853.095400] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 853.095400] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] self.driver.spawn(context, instance, image_meta, [ 853.095400] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 853.095400] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] self._vmops.spawn(context, instance, image_meta, injected_files, [ 853.095400] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 853.095400] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] vm_ref = self.build_virtual_machine(instance, [ 853.095400] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 853.095400] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] vif_infos = vmwarevif.get_vif_info(self._session, [ 853.095400] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 853.095750] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] for vif in network_info: [ 853.095750] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 853.095750] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] return self._sync_wrapper(fn, *args, **kwargs) [ 853.095750] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 853.095750] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] self.wait() [ 853.095750] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 853.095750] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] self[:] = self._gt.wait() [ 853.095750] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 853.095750] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] return self._exit_event.wait() [ 853.095750] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 853.095750] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] result = hub.switch() [ 853.095750] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 853.095750] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] return self.greenlet.switch() [ 853.096236] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 853.096236] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] result = function(*args, **kwargs) [ 853.096236] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 853.096236] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] return func(*args, **kwargs) [ 853.096236] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 853.096236] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] raise e [ 853.096236] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 853.096236] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] nwinfo = self.network_api.allocate_for_instance( [ 853.096236] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 853.096236] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] created_port_ids = self._update_ports_for_instance( [ 853.096236] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 853.096236] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] with excutils.save_and_reraise_exception(): [ 853.096236] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 853.096709] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] self.force_reraise() [ 853.096709] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 853.096709] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] raise self.value [ 853.096709] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 853.096709] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] updated_port = self._update_port( [ 853.096709] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 853.096709] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] _ensure_no_port_binding_failure(port) [ 853.096709] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 853.096709] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] raise exception.PortBindingFailed(port_id=port['id']) [ 853.096709] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] nova.exception.PortBindingFailed: Binding failed for port 018c3cd4-50ad-4134-8ee4-c8aff36ab617, please check neutron logs for more information. [ 853.096709] env[61906]: ERROR nova.compute.manager [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] [ 853.097034] env[61906]: DEBUG nova.compute.utils [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Binding failed for port 018c3cd4-50ad-4134-8ee4-c8aff36ab617, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 853.097461] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.785s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.100544] env[61906]: DEBUG nova.compute.manager [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Build of instance 52a387d2-4090-4eba-9a77-98c37f152b75 was re-scheduled: Binding failed for port 018c3cd4-50ad-4134-8ee4-c8aff36ab617, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 853.100810] env[61906]: DEBUG nova.compute.manager [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 853.102116] env[61906]: DEBUG oslo_concurrency.lockutils [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Acquiring lock "refresh_cache-52a387d2-4090-4eba-9a77-98c37f152b75" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.102116] env[61906]: DEBUG oslo_concurrency.lockutils [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Acquired lock "refresh_cache-52a387d2-4090-4eba-9a77-98c37f152b75" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.102116] env[61906]: DEBUG nova.network.neutron [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 853.177990] env[61906]: DEBUG nova.compute.manager [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 853.302403] env[61906]: DEBUG nova.compute.manager [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 853.326553] env[61906]: DEBUG nova.virt.hardware [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 853.326805] env[61906]: DEBUG nova.virt.hardware [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 853.326963] env[61906]: DEBUG nova.virt.hardware [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 853.327158] env[61906]: DEBUG nova.virt.hardware [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 853.327304] env[61906]: DEBUG nova.virt.hardware [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 853.327448] env[61906]: DEBUG nova.virt.hardware [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 853.327653] env[61906]: DEBUG nova.virt.hardware [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 853.327810] env[61906]: DEBUG nova.virt.hardware [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 853.327973] env[61906]: DEBUG nova.virt.hardware [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 853.328153] env[61906]: DEBUG nova.virt.hardware [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 853.328368] env[61906]: DEBUG nova.virt.hardware [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 853.329252] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56fe16ca-76ba-40b0-9e7d-6a3bdadae920 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.337773] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-468610a0-1fe8-44a9-93d3-ac2a57c9fc3b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.698742] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.732962] env[61906]: DEBUG nova.network.neutron [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 853.827874] env[61906]: DEBUG nova.network.neutron [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.852016] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c95c21d-b00d-4e30-90ae-9f33611451b8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.861924] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b34f9742-c270-4bc5-be3b-423901471f2f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.892034] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e681bacf-963d-4921-ad99-8e275dde2483 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.899289] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9875c41b-c164-4931-9b79-8ab5005387e2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.914774] env[61906]: DEBUG nova.compute.provider_tree [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 854.113587] env[61906]: DEBUG nova.compute.manager [req-9c5c5e7c-b340-496d-815d-90470d56dc9c req-6518f152-8a1f-44ca-9ee7-bc78404a477e service nova] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Received event network-vif-plugged-d5ae18fb-369d-4c5b-9efe-a7d2898521e9 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.113834] env[61906]: DEBUG oslo_concurrency.lockutils [req-9c5c5e7c-b340-496d-815d-90470d56dc9c req-6518f152-8a1f-44ca-9ee7-bc78404a477e service nova] Acquiring lock "7e05fee8-842f-4e03-853a-58bac5a9ac0e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.114278] env[61906]: DEBUG oslo_concurrency.lockutils [req-9c5c5e7c-b340-496d-815d-90470d56dc9c req-6518f152-8a1f-44ca-9ee7-bc78404a477e service nova] Lock "7e05fee8-842f-4e03-853a-58bac5a9ac0e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.114498] env[61906]: DEBUG oslo_concurrency.lockutils [req-9c5c5e7c-b340-496d-815d-90470d56dc9c req-6518f152-8a1f-44ca-9ee7-bc78404a477e service nova] Lock "7e05fee8-842f-4e03-853a-58bac5a9ac0e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.114674] env[61906]: DEBUG nova.compute.manager [req-9c5c5e7c-b340-496d-815d-90470d56dc9c req-6518f152-8a1f-44ca-9ee7-bc78404a477e service nova] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] No waiting events found dispatching network-vif-plugged-d5ae18fb-369d-4c5b-9efe-a7d2898521e9 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 854.114836] env[61906]: WARNING nova.compute.manager [req-9c5c5e7c-b340-496d-815d-90470d56dc9c req-6518f152-8a1f-44ca-9ee7-bc78404a477e service nova] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Received unexpected event network-vif-plugged-d5ae18fb-369d-4c5b-9efe-a7d2898521e9 for instance with vm_state building and task_state spawning. [ 854.204359] env[61906]: DEBUG nova.network.neutron [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Successfully updated port: d5ae18fb-369d-4c5b-9efe-a7d2898521e9 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 854.330404] env[61906]: DEBUG oslo_concurrency.lockutils [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Releasing lock "refresh_cache-52a387d2-4090-4eba-9a77-98c37f152b75" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.330664] env[61906]: DEBUG nova.compute.manager [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 854.330851] env[61906]: DEBUG nova.compute.manager [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 854.331038] env[61906]: DEBUG nova.network.neutron [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 854.346148] env[61906]: DEBUG nova.network.neutron [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 854.417192] env[61906]: DEBUG nova.scheduler.client.report [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 854.849217] env[61906]: DEBUG nova.network.neutron [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.922940] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.824s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.922940] env[61906]: ERROR nova.compute.manager [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f16fa600-42c4-472d-8c4d-468806604d81, please check neutron logs for more information. [ 854.922940] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Traceback (most recent call last): [ 854.922940] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 854.922940] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] self.driver.spawn(context, instance, image_meta, [ 854.922940] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 854.922940] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] self._vmops.spawn(context, instance, image_meta, injected_files, [ 854.922940] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 854.922940] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] vm_ref = self.build_virtual_machine(instance, [ 854.923603] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 854.923603] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] vif_infos = vmwarevif.get_vif_info(self._session, [ 854.923603] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 854.923603] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] for vif in network_info: [ 854.923603] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 854.923603] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] return self._sync_wrapper(fn, *args, **kwargs) [ 854.923603] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 854.923603] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] self.wait() [ 854.923603] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 854.923603] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] self[:] = self._gt.wait() [ 854.923603] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 854.923603] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] return self._exit_event.wait() [ 854.923603] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 854.924123] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] result = hub.switch() [ 854.924123] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 854.924123] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] return self.greenlet.switch() [ 854.924123] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 854.924123] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] result = function(*args, **kwargs) [ 854.924123] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 854.924123] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] return func(*args, **kwargs) [ 854.924123] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 854.924123] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] raise e [ 854.924123] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 854.924123] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] nwinfo = self.network_api.allocate_for_instance( [ 854.924123] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 854.924123] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] created_port_ids = self._update_ports_for_instance( [ 854.924961] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 854.924961] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] with excutils.save_and_reraise_exception(): [ 854.924961] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 854.924961] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] self.force_reraise() [ 854.924961] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 854.924961] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] raise self.value [ 854.924961] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 854.924961] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] updated_port = self._update_port( [ 854.924961] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 854.924961] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] _ensure_no_port_binding_failure(port) [ 854.924961] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 854.924961] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] raise exception.PortBindingFailed(port_id=port['id']) [ 854.925372] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] nova.exception.PortBindingFailed: Binding failed for port f16fa600-42c4-472d-8c4d-468806604d81, please check neutron logs for more information. [ 854.925372] env[61906]: ERROR nova.compute.manager [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] [ 854.925372] env[61906]: DEBUG nova.compute.utils [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Binding failed for port f16fa600-42c4-472d-8c4d-468806604d81, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 854.925372] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.059s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.928616] env[61906]: DEBUG nova.compute.manager [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Build of instance 34f333be-ea19-499e-ab89-e9f7663ae596 was re-scheduled: Binding failed for port f16fa600-42c4-472d-8c4d-468806604d81, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 854.929054] env[61906]: DEBUG nova.compute.manager [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 854.929289] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Acquiring lock "refresh_cache-34f333be-ea19-499e-ab89-e9f7663ae596" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.929437] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Acquired lock "refresh_cache-34f333be-ea19-499e-ab89-e9f7663ae596" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.929681] env[61906]: DEBUG nova.network.neutron [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 855.356420] env[61906]: INFO nova.compute.manager [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] [instance: 52a387d2-4090-4eba-9a77-98c37f152b75] Took 1.03 seconds to deallocate network for instance. [ 855.449577] env[61906]: DEBUG nova.network.neutron [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 855.516229] env[61906]: DEBUG nova.network.neutron [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.667721] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d05b0f3a-1c5c-40c9-a081-c00caccb26ff {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.676034] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a19719-8cbb-4541-b871-b0faa13b9855 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.706459] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3705eb6e-9026-4f6a-ac1c-3f97f2ddab7b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.714161] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc182386-5673-4b52-b3ac-0381236582b0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.726842] env[61906]: DEBUG nova.compute.provider_tree [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 856.019303] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Releasing lock "refresh_cache-34f333be-ea19-499e-ab89-e9f7663ae596" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.019583] env[61906]: DEBUG nova.compute.manager [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 856.019781] env[61906]: DEBUG nova.compute.manager [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 856.019998] env[61906]: DEBUG nova.network.neutron [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 856.034958] env[61906]: DEBUG nova.network.neutron [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 856.145644] env[61906]: DEBUG nova.compute.manager [req-b2b143d0-9299-4d8f-9752-2758c53adccb req-497a0a68-d422-44a1-ba35-adf0cb9431eb service nova] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Received event network-changed-d5ae18fb-369d-4c5b-9efe-a7d2898521e9 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 856.145863] env[61906]: DEBUG nova.compute.manager [req-b2b143d0-9299-4d8f-9752-2758c53adccb req-497a0a68-d422-44a1-ba35-adf0cb9431eb service nova] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Refreshing instance network info cache due to event network-changed-d5ae18fb-369d-4c5b-9efe-a7d2898521e9. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 856.147786] env[61906]: DEBUG oslo_concurrency.lockutils [req-b2b143d0-9299-4d8f-9752-2758c53adccb req-497a0a68-d422-44a1-ba35-adf0cb9431eb service nova] Acquiring lock "refresh_cache-7e05fee8-842f-4e03-853a-58bac5a9ac0e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.148468] env[61906]: DEBUG oslo_concurrency.lockutils [req-b2b143d0-9299-4d8f-9752-2758c53adccb req-497a0a68-d422-44a1-ba35-adf0cb9431eb service nova] Acquired lock "refresh_cache-7e05fee8-842f-4e03-853a-58bac5a9ac0e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.148468] env[61906]: DEBUG nova.network.neutron [req-b2b143d0-9299-4d8f-9752-2758c53adccb req-497a0a68-d422-44a1-ba35-adf0cb9431eb service nova] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Refreshing network info cache for port d5ae18fb-369d-4c5b-9efe-a7d2898521e9 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 856.229950] env[61906]: DEBUG nova.scheduler.client.report [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 856.394343] env[61906]: INFO nova.scheduler.client.report [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Deleted allocations for instance 52a387d2-4090-4eba-9a77-98c37f152b75 [ 856.531305] env[61906]: DEBUG nova.network.neutron [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Successfully updated port: c156b7ea-028a-4b34-93cd-25d384ab5e76 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 856.537542] env[61906]: DEBUG nova.network.neutron [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.683256] env[61906]: DEBUG nova.network.neutron [req-b2b143d0-9299-4d8f-9752-2758c53adccb req-497a0a68-d422-44a1-ba35-adf0cb9431eb service nova] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 856.735720] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.811s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.736370] env[61906]: ERROR nova.compute.manager [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3644c245-2b19-4580-b700-95677871b770, please check neutron logs for more information. [ 856.736370] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Traceback (most recent call last): [ 856.736370] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 856.736370] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] self.driver.spawn(context, instance, image_meta, [ 856.736370] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 856.736370] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 856.736370] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 856.736370] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] vm_ref = self.build_virtual_machine(instance, [ 856.736370] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 856.736370] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] vif_infos = vmwarevif.get_vif_info(self._session, [ 856.736370] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 856.736661] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] for vif in network_info: [ 856.736661] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 856.736661] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] return self._sync_wrapper(fn, *args, **kwargs) [ 856.736661] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 856.736661] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] self.wait() [ 856.736661] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 856.736661] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] self[:] = self._gt.wait() [ 856.736661] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 856.736661] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] return self._exit_event.wait() [ 856.736661] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 856.736661] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] result = hub.switch() [ 856.736661] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 856.736661] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] return self.greenlet.switch() [ 856.737015] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 856.737015] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] result = function(*args, **kwargs) [ 856.737015] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 856.737015] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] return func(*args, **kwargs) [ 856.737015] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 856.737015] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] raise e [ 856.737015] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 856.737015] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] nwinfo = self.network_api.allocate_for_instance( [ 856.737015] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 856.737015] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] created_port_ids = self._update_ports_for_instance( [ 856.737015] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 856.737015] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] with excutils.save_and_reraise_exception(): [ 856.737015] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 856.737385] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] self.force_reraise() [ 856.737385] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 856.737385] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] raise self.value [ 856.737385] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 856.737385] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] updated_port = self._update_port( [ 856.737385] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 856.737385] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] _ensure_no_port_binding_failure(port) [ 856.737385] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 856.737385] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] raise exception.PortBindingFailed(port_id=port['id']) [ 856.737385] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] nova.exception.PortBindingFailed: Binding failed for port 3644c245-2b19-4580-b700-95677871b770, please check neutron logs for more information. [ 856.737385] env[61906]: ERROR nova.compute.manager [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] [ 856.737639] env[61906]: DEBUG nova.compute.utils [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Binding failed for port 3644c245-2b19-4580-b700-95677871b770, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 856.738760] env[61906]: DEBUG oslo_concurrency.lockutils [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.541s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.742452] env[61906]: DEBUG nova.compute.manager [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Build of instance 7183e141-9989-4a90-b2fb-2ca0d8e6cda9 was re-scheduled: Binding failed for port 3644c245-2b19-4580-b700-95677871b770, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 856.742977] env[61906]: DEBUG nova.compute.manager [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 856.743220] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Acquiring lock "refresh_cache-7183e141-9989-4a90-b2fb-2ca0d8e6cda9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.743368] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Acquired lock "refresh_cache-7183e141-9989-4a90-b2fb-2ca0d8e6cda9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.743530] env[61906]: DEBUG nova.network.neutron [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 856.758238] env[61906]: DEBUG nova.network.neutron [req-b2b143d0-9299-4d8f-9752-2758c53adccb req-497a0a68-d422-44a1-ba35-adf0cb9431eb service nova] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.908026] env[61906]: DEBUG oslo_concurrency.lockutils [None req-163be0b0-83d5-46f7-9853-20eda0b46f5c tempest-ServerMetadataNegativeTestJSON-164917139 tempest-ServerMetadataNegativeTestJSON-164917139-project-member] Lock "52a387d2-4090-4eba-9a77-98c37f152b75" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.887s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.033851] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Acquiring lock "refresh_cache-7e05fee8-842f-4e03-853a-58bac5a9ac0e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.040687] env[61906]: INFO nova.compute.manager [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] [instance: 34f333be-ea19-499e-ab89-e9f7663ae596] Took 1.02 seconds to deallocate network for instance. [ 857.261013] env[61906]: DEBUG oslo_concurrency.lockutils [req-b2b143d0-9299-4d8f-9752-2758c53adccb req-497a0a68-d422-44a1-ba35-adf0cb9431eb service nova] Releasing lock "refresh_cache-7e05fee8-842f-4e03-853a-58bac5a9ac0e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.261741] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Acquired lock "refresh_cache-7e05fee8-842f-4e03-853a-58bac5a9ac0e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.261741] env[61906]: DEBUG nova.network.neutron [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 857.278936] env[61906]: DEBUG nova.network.neutron [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 857.377496] env[61906]: DEBUG nova.network.neutron [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.418504] env[61906]: DEBUG nova.compute.manager [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 857.524026] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb15b59c-5748-42c5-a6e9-56f36f9c9a05 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.535764] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d1891b4-8518-4c33-9339-947e7ae36d24 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.571403] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4855a1e-632f-434e-9cbf-df3ec2bcaf4e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.580797] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19626c66-9dcd-4ac5-aad3-b7030235bc44 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.608989] env[61906]: DEBUG nova.compute.provider_tree [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.806664] env[61906]: DEBUG nova.network.neutron [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 857.883521] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Releasing lock "refresh_cache-7183e141-9989-4a90-b2fb-2ca0d8e6cda9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.883521] env[61906]: DEBUG nova.compute.manager [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 857.883521] env[61906]: DEBUG nova.compute.manager [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 857.883701] env[61906]: DEBUG nova.network.neutron [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 857.935758] env[61906]: DEBUG nova.network.neutron [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 857.948111] env[61906]: DEBUG oslo_concurrency.lockutils [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.112428] env[61906]: INFO nova.scheduler.client.report [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Deleted allocations for instance 34f333be-ea19-499e-ab89-e9f7663ae596 [ 858.126263] env[61906]: DEBUG nova.scheduler.client.report [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 858.241978] env[61906]: DEBUG nova.compute.manager [req-75e81998-10bd-4c91-a7a5-8501c358f32b req-144c180d-6730-424d-8612-763916d65233 service nova] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Received event network-vif-plugged-c156b7ea-028a-4b34-93cd-25d384ab5e76 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 858.242261] env[61906]: DEBUG oslo_concurrency.lockutils [req-75e81998-10bd-4c91-a7a5-8501c358f32b req-144c180d-6730-424d-8612-763916d65233 service nova] Acquiring lock "7e05fee8-842f-4e03-853a-58bac5a9ac0e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.242489] env[61906]: DEBUG oslo_concurrency.lockutils [req-75e81998-10bd-4c91-a7a5-8501c358f32b req-144c180d-6730-424d-8612-763916d65233 service nova] Lock "7e05fee8-842f-4e03-853a-58bac5a9ac0e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.242702] env[61906]: DEBUG oslo_concurrency.lockutils [req-75e81998-10bd-4c91-a7a5-8501c358f32b req-144c180d-6730-424d-8612-763916d65233 service nova] Lock "7e05fee8-842f-4e03-853a-58bac5a9ac0e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.242886] env[61906]: DEBUG nova.compute.manager [req-75e81998-10bd-4c91-a7a5-8501c358f32b req-144c180d-6730-424d-8612-763916d65233 service nova] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] No waiting events found dispatching network-vif-plugged-c156b7ea-028a-4b34-93cd-25d384ab5e76 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 858.243557] env[61906]: WARNING nova.compute.manager [req-75e81998-10bd-4c91-a7a5-8501c358f32b req-144c180d-6730-424d-8612-763916d65233 service nova] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Received unexpected event network-vif-plugged-c156b7ea-028a-4b34-93cd-25d384ab5e76 for instance with vm_state building and task_state spawning. [ 858.243642] env[61906]: DEBUG nova.compute.manager [req-75e81998-10bd-4c91-a7a5-8501c358f32b req-144c180d-6730-424d-8612-763916d65233 service nova] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Received event network-changed-c156b7ea-028a-4b34-93cd-25d384ab5e76 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 858.243798] env[61906]: DEBUG nova.compute.manager [req-75e81998-10bd-4c91-a7a5-8501c358f32b req-144c180d-6730-424d-8612-763916d65233 service nova] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Refreshing instance network info cache due to event network-changed-c156b7ea-028a-4b34-93cd-25d384ab5e76. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 858.244538] env[61906]: DEBUG oslo_concurrency.lockutils [req-75e81998-10bd-4c91-a7a5-8501c358f32b req-144c180d-6730-424d-8612-763916d65233 service nova] Acquiring lock "refresh_cache-7e05fee8-842f-4e03-853a-58bac5a9ac0e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.383381] env[61906]: DEBUG nova.network.neutron [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Updating instance_info_cache with network_info: [{"id": "d5ae18fb-369d-4c5b-9efe-a7d2898521e9", "address": "fa:16:3e:fd:f3:a3", "network": {"id": "b97b90a3-e246-4b73-8179-e0a383ae9e1e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1071998974", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.108", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c15bc344a5440d491b47b9fdedb6869", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5ae18fb-36", "ovs_interfaceid": "d5ae18fb-369d-4c5b-9efe-a7d2898521e9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c156b7ea-028a-4b34-93cd-25d384ab5e76", "address": "fa:16:3e:a3:04:57", "network": {"id": "3b4f4ff7-0e0c-47f3-8bba-0f97f597f8ec", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-993107218", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.143", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4c15bc344a5440d491b47b9fdedb6869", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc156b7ea-02", "ovs_interfaceid": "c156b7ea-028a-4b34-93cd-25d384ab5e76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.438993] env[61906]: DEBUG nova.network.neutron [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.638988] env[61906]: DEBUG oslo_concurrency.lockutils [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.900s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.639689] env[61906]: ERROR nova.compute.manager [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7033b729-91d5-4497-adeb-aea522b40942, please check neutron logs for more information. [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Traceback (most recent call last): [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] self.driver.spawn(context, instance, image_meta, [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] vm_ref = self.build_virtual_machine(instance, [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] vif_infos = vmwarevif.get_vif_info(self._session, [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] for vif in network_info: [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] return self._sync_wrapper(fn, *args, **kwargs) [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] self.wait() [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] self[:] = self._gt.wait() [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] return self._exit_event.wait() [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] result = hub.switch() [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] return self.greenlet.switch() [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] result = function(*args, **kwargs) [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] return func(*args, **kwargs) [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] raise e [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] nwinfo = self.network_api.allocate_for_instance( [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] created_port_ids = self._update_ports_for_instance( [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] with excutils.save_and_reraise_exception(): [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 858.639689] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] self.force_reraise() [ 858.641244] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 858.641244] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] raise self.value [ 858.641244] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 858.641244] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] updated_port = self._update_port( [ 858.641244] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 858.641244] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] _ensure_no_port_binding_failure(port) [ 858.641244] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 858.641244] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] raise exception.PortBindingFailed(port_id=port['id']) [ 858.641244] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] nova.exception.PortBindingFailed: Binding failed for port 7033b729-91d5-4497-adeb-aea522b40942, please check neutron logs for more information. [ 858.641244] env[61906]: ERROR nova.compute.manager [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] [ 858.641244] env[61906]: DEBUG nova.compute.utils [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Binding failed for port 7033b729-91d5-4497-adeb-aea522b40942, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 858.642148] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.115s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.646466] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a2183946-bd7d-4201-9f9d-242b9b1a5e08 tempest-ServersAdminNegativeTestJSON-1984475886 tempest-ServersAdminNegativeTestJSON-1984475886-project-member] Lock "34f333be-ea19-499e-ab89-e9f7663ae596" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.102s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.646466] env[61906]: DEBUG nova.compute.manager [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Build of instance e0afe101-f4a4-4dc8-a666-9a7793dbf8a6 was re-scheduled: Binding failed for port 7033b729-91d5-4497-adeb-aea522b40942, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 858.646466] env[61906]: DEBUG nova.compute.manager [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 858.646466] env[61906]: DEBUG oslo_concurrency.lockutils [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "refresh_cache-e0afe101-f4a4-4dc8-a666-9a7793dbf8a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.646466] env[61906]: DEBUG oslo_concurrency.lockutils [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquired lock "refresh_cache-e0afe101-f4a4-4dc8-a666-9a7793dbf8a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.646466] env[61906]: DEBUG nova.network.neutron [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 858.887337] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Releasing lock "refresh_cache-7e05fee8-842f-4e03-853a-58bac5a9ac0e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.887337] env[61906]: DEBUG nova.compute.manager [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Instance network_info: |[{"id": "d5ae18fb-369d-4c5b-9efe-a7d2898521e9", "address": "fa:16:3e:fd:f3:a3", "network": {"id": "b97b90a3-e246-4b73-8179-e0a383ae9e1e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1071998974", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.108", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c15bc344a5440d491b47b9fdedb6869", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5ae18fb-36", "ovs_interfaceid": "d5ae18fb-369d-4c5b-9efe-a7d2898521e9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c156b7ea-028a-4b34-93cd-25d384ab5e76", "address": "fa:16:3e:a3:04:57", "network": {"id": "3b4f4ff7-0e0c-47f3-8bba-0f97f597f8ec", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-993107218", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.143", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4c15bc344a5440d491b47b9fdedb6869", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc156b7ea-02", "ovs_interfaceid": "c156b7ea-028a-4b34-93cd-25d384ab5e76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 858.887337] env[61906]: DEBUG oslo_concurrency.lockutils [req-75e81998-10bd-4c91-a7a5-8501c358f32b req-144c180d-6730-424d-8612-763916d65233 service nova] Acquired lock "refresh_cache-7e05fee8-842f-4e03-853a-58bac5a9ac0e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.887337] env[61906]: DEBUG nova.network.neutron [req-75e81998-10bd-4c91-a7a5-8501c358f32b req-144c180d-6730-424d-8612-763916d65233 service nova] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Refreshing network info cache for port c156b7ea-028a-4b34-93cd-25d384ab5e76 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 858.890605] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fd:f3:a3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3734b156-0f7d-4721-b23c-d000412ec2eb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd5ae18fb-369d-4c5b-9efe-a7d2898521e9', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:a3:04:57', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb224918-e294-4b76-80f9-2fa0031b7dc2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c156b7ea-028a-4b34-93cd-25d384ab5e76', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 858.903844] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Creating folder: Project (4c15bc344a5440d491b47b9fdedb6869). Parent ref: group-v284713. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 858.907989] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d1ca133b-1e07-47dd-861f-d25276737ea6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.925549] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Created folder: Project (4c15bc344a5440d491b47b9fdedb6869) in parent group-v284713. [ 858.925760] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Creating folder: Instances. Parent ref: group-v284738. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 858.926011] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c6e239f1-61e2-4415-9d82-213e27a0927a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.937504] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Created folder: Instances in parent group-v284738. [ 858.937670] env[61906]: DEBUG oslo.service.loopingcall [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 858.938151] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 858.938397] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-93a242f1-4987-4d15-926e-1c8274b2de1d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.959118] env[61906]: INFO nova.compute.manager [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] [instance: 7183e141-9989-4a90-b2fb-2ca0d8e6cda9] Took 1.08 seconds to deallocate network for instance. [ 858.968692] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 858.968692] env[61906]: value = "task-1333169" [ 858.968692] env[61906]: _type = "Task" [ 858.968692] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.980252] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333169, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.151620] env[61906]: DEBUG nova.compute.manager [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 859.178327] env[61906]: DEBUG nova.network.neutron [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 859.348822] env[61906]: DEBUG nova.network.neutron [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.415659] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1145ca5-64c8-4a8d-a0d1-70e4dca730da {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.423844] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e5e6568-86f8-462a-ae40-efa08c99125e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.464212] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92c9a4f9-3744-4f32-8e5e-65914a066897 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.480875] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7327056-2e21-4fed-99e4-56770b308616 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.491360] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333169, 'name': CreateVM_Task, 'duration_secs': 0.388824} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.491999] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 859.502923] env[61906]: DEBUG nova.compute.provider_tree [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 859.509488] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.509899] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.510391] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 859.510739] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93a64341-5b7d-4d53-b39f-4b6f86d0aaa8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.517167] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Waiting for the task: (returnval){ [ 859.517167] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52221c3c-d732-4866-0d34-342fa4cc9955" [ 859.517167] env[61906]: _type = "Task" [ 859.517167] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.527607] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52221c3c-d732-4866-0d34-342fa4cc9955, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.677078] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.851907] env[61906]: DEBUG oslo_concurrency.lockutils [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Releasing lock "refresh_cache-e0afe101-f4a4-4dc8-a666-9a7793dbf8a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.852199] env[61906]: DEBUG nova.compute.manager [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 859.852390] env[61906]: DEBUG nova.compute.manager [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 859.852560] env[61906]: DEBUG nova.network.neutron [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 859.879483] env[61906]: DEBUG nova.network.neutron [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 859.992130] env[61906]: DEBUG nova.network.neutron [req-75e81998-10bd-4c91-a7a5-8501c358f32b req-144c180d-6730-424d-8612-763916d65233 service nova] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Updated VIF entry in instance network info cache for port c156b7ea-028a-4b34-93cd-25d384ab5e76. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 859.992130] env[61906]: DEBUG nova.network.neutron [req-75e81998-10bd-4c91-a7a5-8501c358f32b req-144c180d-6730-424d-8612-763916d65233 service nova] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Updating instance_info_cache with network_info: [{"id": "d5ae18fb-369d-4c5b-9efe-a7d2898521e9", "address": "fa:16:3e:fd:f3:a3", "network": {"id": "b97b90a3-e246-4b73-8179-e0a383ae9e1e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1071998974", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.108", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c15bc344a5440d491b47b9fdedb6869", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5ae18fb-36", "ovs_interfaceid": "d5ae18fb-369d-4c5b-9efe-a7d2898521e9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c156b7ea-028a-4b34-93cd-25d384ab5e76", "address": "fa:16:3e:a3:04:57", "network": {"id": "3b4f4ff7-0e0c-47f3-8bba-0f97f597f8ec", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-993107218", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.143", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4c15bc344a5440d491b47b9fdedb6869", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc156b7ea-02", "ovs_interfaceid": "c156b7ea-028a-4b34-93cd-25d384ab5e76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.006331] env[61906]: INFO nova.scheduler.client.report [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Deleted allocations for instance 7183e141-9989-4a90-b2fb-2ca0d8e6cda9 [ 860.015896] env[61906]: DEBUG nova.scheduler.client.report [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 860.041618] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52221c3c-d732-4866-0d34-342fa4cc9955, 'name': SearchDatastore_Task, 'duration_secs': 0.013502} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.041977] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.042217] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 860.043903] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.043903] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.043903] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 860.043903] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-709deb4b-c596-46ff-b1f1-0497fe624c9a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.052943] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 860.053764] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 860.054190] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-976bb051-cf16-4982-a1ef-15dcb81e290e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.062213] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Waiting for the task: (returnval){ [ 860.062213] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52df9aff-8519-c261-c355-810aa88575aa" [ 860.062213] env[61906]: _type = "Task" [ 860.062213] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.072792] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52df9aff-8519-c261-c355-810aa88575aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.382307] env[61906]: DEBUG nova.network.neutron [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.390747] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "b1159533-c970-49d6-af42-b954b20d92fb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.390982] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "b1159533-c970-49d6-af42-b954b20d92fb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.493818] env[61906]: DEBUG oslo_concurrency.lockutils [req-75e81998-10bd-4c91-a7a5-8501c358f32b req-144c180d-6730-424d-8612-763916d65233 service nova] Releasing lock "refresh_cache-7e05fee8-842f-4e03-853a-58bac5a9ac0e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.525869] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.884s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.526595] env[61906]: ERROR nova.compute.manager [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 122dafdd-c10c-4440-84d5-755b9239a2ce, please check neutron logs for more information. [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] Traceback (most recent call last): [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] self.driver.spawn(context, instance, image_meta, [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] self._vmops.spawn(context, instance, image_meta, injected_files, [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] vm_ref = self.build_virtual_machine(instance, [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] vif_infos = vmwarevif.get_vif_info(self._session, [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] for vif in network_info: [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] return self._sync_wrapper(fn, *args, **kwargs) [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] self.wait() [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] self[:] = self._gt.wait() [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] return self._exit_event.wait() [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] current.throw(*self._exc) [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] result = function(*args, **kwargs) [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] return func(*args, **kwargs) [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] raise e [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] nwinfo = self.network_api.allocate_for_instance( [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] created_port_ids = self._update_ports_for_instance( [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] with excutils.save_and_reraise_exception(): [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] self.force_reraise() [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 860.526595] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] raise self.value [ 860.527630] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 860.527630] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] updated_port = self._update_port( [ 860.527630] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 860.527630] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] _ensure_no_port_binding_failure(port) [ 860.527630] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 860.527630] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] raise exception.PortBindingFailed(port_id=port['id']) [ 860.527630] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] nova.exception.PortBindingFailed: Binding failed for port 122dafdd-c10c-4440-84d5-755b9239a2ce, please check neutron logs for more information. [ 860.527630] env[61906]: ERROR nova.compute.manager [instance: a651bd36-b527-4743-bc6e-5534c629e409] [ 860.527630] env[61906]: DEBUG nova.compute.utils [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Binding failed for port 122dafdd-c10c-4440-84d5-755b9239a2ce, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 860.529173] env[61906]: DEBUG nova.compute.manager [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Build of instance a651bd36-b527-4743-bc6e-5534c629e409 was re-scheduled: Binding failed for port 122dafdd-c10c-4440-84d5-755b9239a2ce, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 860.529626] env[61906]: DEBUG nova.compute.manager [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 860.529895] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Acquiring lock "refresh_cache-a651bd36-b527-4743-bc6e-5534c629e409" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.530091] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Acquired lock "refresh_cache-a651bd36-b527-4743-bc6e-5534c629e409" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.530350] env[61906]: DEBUG nova.network.neutron [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 860.531428] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.643s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.535127] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6e00e706-1de0-4cd6-a1ee-ff40ee66b051 tempest-ServerAddressesTestJSON-1643199274 tempest-ServerAddressesTestJSON-1643199274-project-member] Lock "7183e141-9989-4a90-b2fb-2ca0d8e6cda9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.777s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.574737] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52df9aff-8519-c261-c355-810aa88575aa, 'name': SearchDatastore_Task, 'duration_secs': 0.00908} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.575800] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1532ba8e-c942-4272-8822-a53490813cea {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.584383] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Waiting for the task: (returnval){ [ 860.584383] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5241b752-7c2a-c61e-23af-49e2443e382d" [ 860.584383] env[61906]: _type = "Task" [ 860.584383] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.592551] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5241b752-7c2a-c61e-23af-49e2443e382d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.885553] env[61906]: INFO nova.compute.manager [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: e0afe101-f4a4-4dc8-a666-9a7793dbf8a6] Took 1.03 seconds to deallocate network for instance. [ 861.042240] env[61906]: DEBUG nova.compute.manager [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 861.058839] env[61906]: DEBUG nova.network.neutron [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 861.095737] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5241b752-7c2a-c61e-23af-49e2443e382d, 'name': SearchDatastore_Task, 'duration_secs': 0.009201} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.098285] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.098547] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 7e05fee8-842f-4e03-853a-58bac5a9ac0e/7e05fee8-842f-4e03-853a-58bac5a9ac0e.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 861.099725] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5e1d77a4-bc33-4529-ac6b-888cba7f9bbb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.106945] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Waiting for the task: (returnval){ [ 861.106945] env[61906]: value = "task-1333170" [ 861.106945] env[61906]: _type = "Task" [ 861.106945] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.118696] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Task: {'id': task-1333170, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.168351] env[61906]: DEBUG nova.network.neutron [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.252446] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47985d88-1a8c-4971-8e5d-7f7150b1ec24 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.260745] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74755ede-b5dc-442c-adad-722230d7c4f7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.292955] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a8f8ef8-9078-4e29-9f29-5e388b9f346e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.301959] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4d4d633-790f-48df-a00b-2d68196c4dc0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.317857] env[61906]: DEBUG nova.compute.provider_tree [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 861.385203] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 861.385997] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 861.562769] env[61906]: DEBUG oslo_concurrency.lockutils [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.618124] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Task: {'id': task-1333170, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.676422] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Releasing lock "refresh_cache-a651bd36-b527-4743-bc6e-5534c629e409" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.676422] env[61906]: DEBUG nova.compute.manager [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 861.676422] env[61906]: DEBUG nova.compute.manager [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 861.676422] env[61906]: DEBUG nova.network.neutron [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 861.689993] env[61906]: DEBUG nova.network.neutron [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 861.820883] env[61906]: DEBUG nova.scheduler.client.report [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 861.917955] env[61906]: INFO nova.scheduler.client.report [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Deleted allocations for instance e0afe101-f4a4-4dc8-a666-9a7793dbf8a6 [ 862.122021] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Task: {'id': task-1333170, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.525568} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.122021] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 7e05fee8-842f-4e03-853a-58bac5a9ac0e/7e05fee8-842f-4e03-853a-58bac5a9ac0e.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 862.122021] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 862.122264] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7a688df0-1553-4b1a-8727-c705c5fb83a4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.130349] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Waiting for the task: (returnval){ [ 862.130349] env[61906]: value = "task-1333171" [ 862.130349] env[61906]: _type = "Task" [ 862.130349] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.139784] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Task: {'id': task-1333171, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.192488] env[61906]: DEBUG nova.network.neutron [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.326736] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.795s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.327418] env[61906]: ERROR nova.compute.manager [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 21804744-845d-412e-b7c7-84b67dbbb153, please check neutron logs for more information. [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Traceback (most recent call last): [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] self.driver.spawn(context, instance, image_meta, [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] vm_ref = self.build_virtual_machine(instance, [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] vif_infos = vmwarevif.get_vif_info(self._session, [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] for vif in network_info: [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] return self._sync_wrapper(fn, *args, **kwargs) [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] self.wait() [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] self[:] = self._gt.wait() [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] return self._exit_event.wait() [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] current.throw(*self._exc) [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] result = function(*args, **kwargs) [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] return func(*args, **kwargs) [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] raise e [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] nwinfo = self.network_api.allocate_for_instance( [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] created_port_ids = self._update_ports_for_instance( [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] with excutils.save_and_reraise_exception(): [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] self.force_reraise() [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 862.327418] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] raise self.value [ 862.328811] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 862.328811] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] updated_port = self._update_port( [ 862.328811] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 862.328811] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] _ensure_no_port_binding_failure(port) [ 862.328811] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 862.328811] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] raise exception.PortBindingFailed(port_id=port['id']) [ 862.328811] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] nova.exception.PortBindingFailed: Binding failed for port 21804744-845d-412e-b7c7-84b67dbbb153, please check neutron logs for more information. [ 862.328811] env[61906]: ERROR nova.compute.manager [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] [ 862.328811] env[61906]: DEBUG nova.compute.utils [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Binding failed for port 21804744-845d-412e-b7c7-84b67dbbb153, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 862.329263] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.144s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.330999] env[61906]: INFO nova.compute.claims [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 862.334697] env[61906]: DEBUG nova.compute.manager [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Build of instance 976021e6-0e38-4295-bdd4-83f00a7b7be0 was re-scheduled: Binding failed for port 21804744-845d-412e-b7c7-84b67dbbb153, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 862.336037] env[61906]: DEBUG nova.compute.manager [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 862.336037] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Acquiring lock "refresh_cache-976021e6-0e38-4295-bdd4-83f00a7b7be0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.336037] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Acquired lock "refresh_cache-976021e6-0e38-4295-bdd4-83f00a7b7be0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.336037] env[61906]: DEBUG nova.network.neutron [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 862.387559] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 862.387703] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61906) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 862.387888] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 862.428653] env[61906]: DEBUG oslo_concurrency.lockutils [None req-639182df-b0b6-4ba7-8a34-0cbc3908b707 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "e0afe101-f4a4-4dc8-a666-9a7793dbf8a6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.495s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.643713] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Task: {'id': task-1333171, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062156} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.643713] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 862.644030] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c973de2b-e22f-4e42-9049-470f794c92c2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.675406] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] 7e05fee8-842f-4e03-853a-58bac5a9ac0e/7e05fee8-842f-4e03-853a-58bac5a9ac0e.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 862.675778] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b768948e-f1de-4951-886d-65283b491264 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.695126] env[61906]: INFO nova.compute.manager [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] [instance: a651bd36-b527-4743-bc6e-5534c629e409] Took 1.02 seconds to deallocate network for instance. [ 862.701152] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Waiting for the task: (returnval){ [ 862.701152] env[61906]: value = "task-1333172" [ 862.701152] env[61906]: _type = "Task" [ 862.701152] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.709518] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Task: {'id': task-1333172, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.855210] env[61906]: DEBUG nova.network.neutron [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 862.892399] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.934784] env[61906]: DEBUG nova.compute.manager [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 862.979987] env[61906]: DEBUG nova.network.neutron [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.213938] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Task: {'id': task-1333172, 'name': ReconfigVM_Task, 'duration_secs': 0.269537} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.214262] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Reconfigured VM instance instance-0000003d to attach disk [datastore2] 7e05fee8-842f-4e03-853a-58bac5a9ac0e/7e05fee8-842f-4e03-853a-58bac5a9ac0e.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 863.214944] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7db76359-05c3-48f6-98b0-5b5077d55e5f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.221065] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Waiting for the task: (returnval){ [ 863.221065] env[61906]: value = "task-1333173" [ 863.221065] env[61906]: _type = "Task" [ 863.221065] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.229939] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Task: {'id': task-1333173, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.454613] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.483038] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Releasing lock "refresh_cache-976021e6-0e38-4295-bdd4-83f00a7b7be0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.483296] env[61906]: DEBUG nova.compute.manager [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 863.483480] env[61906]: DEBUG nova.compute.manager [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 863.483660] env[61906]: DEBUG nova.network.neutron [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 863.503888] env[61906]: DEBUG nova.network.neutron [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 863.541762] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50e1cdaf-bc8b-4744-974b-88b557d04946 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.549978] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd4600e3-4e26-42ed-8352-344d4e1cc1ec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.579423] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dec4ae0f-36a5-4015-8f10-b4787c0e50fc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.587609] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c671df0-1826-4538-9f05-98c5e7e43d22 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.600812] env[61906]: DEBUG nova.compute.provider_tree [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.662352] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "1bc9f3d4-b539-4841-b5b6-eaefa0d06a72" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.663037] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "1bc9f3d4-b539-4841-b5b6-eaefa0d06a72" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.730828] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Task: {'id': task-1333173, 'name': Rename_Task, 'duration_secs': 0.134578} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.731800] env[61906]: INFO nova.scheduler.client.report [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Deleted allocations for instance a651bd36-b527-4743-bc6e-5534c629e409 [ 863.737041] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 863.738203] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6e0c62e3-40b1-4ca6-be55-1c3831de89ac {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.745009] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Waiting for the task: (returnval){ [ 863.745009] env[61906]: value = "task-1333174" [ 863.745009] env[61906]: _type = "Task" [ 863.745009] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.759718] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Task: {'id': task-1333174, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.008531] env[61906]: DEBUG nova.network.neutron [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.106159] env[61906]: DEBUG nova.scheduler.client.report [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 864.241324] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5b43888a-8213-466e-8c64-35e6b4609599 tempest-ServersNegativeTestJSON-502207932 tempest-ServersNegativeTestJSON-502207932-project-member] Lock "a651bd36-b527-4743-bc6e-5534c629e409" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.078s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.260339] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Task: {'id': task-1333174, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.514942] env[61906]: INFO nova.compute.manager [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] [instance: 976021e6-0e38-4295-bdd4-83f00a7b7be0] Took 1.03 seconds to deallocate network for instance. [ 864.617176] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.288s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.618050] env[61906]: DEBUG nova.compute.manager [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 864.624148] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.579s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.624148] env[61906]: INFO nova.compute.claims [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 864.744097] env[61906]: DEBUG nova.compute.manager [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 864.763824] env[61906]: DEBUG oslo_vmware.api [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Task: {'id': task-1333174, 'name': PowerOnVM_Task, 'duration_secs': 0.525128} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.764169] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 864.764424] env[61906]: INFO nova.compute.manager [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Took 11.46 seconds to spawn the instance on the hypervisor. [ 864.768028] env[61906]: DEBUG nova.compute.manager [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 864.768028] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ba9ee83-cd79-4b81-9c84-b5505210c276 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.127955] env[61906]: DEBUG nova.compute.utils [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 865.133904] env[61906]: DEBUG nova.compute.manager [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 865.133904] env[61906]: DEBUG nova.network.neutron [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 865.243426] env[61906]: DEBUG nova.policy [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bfb8ebd8f94f45dfadb1d8802d04aec9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4f7fc2042e6549cabe98838990f35842', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 865.290261] env[61906]: INFO nova.compute.manager [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Took 41.39 seconds to build instance. [ 865.293675] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.568169] env[61906]: INFO nova.scheduler.client.report [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Deleted allocations for instance 976021e6-0e38-4295-bdd4-83f00a7b7be0 [ 865.638066] env[61906]: DEBUG nova.compute.manager [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 865.735302] env[61906]: DEBUG nova.network.neutron [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Successfully created port: 61a42623-e2e5-4245-9560-be57af9a8397 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 865.793594] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d4ecfab7-7d19-47a5-810e-252a4b1575c2 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Lock "7e05fee8-842f-4e03-853a-58bac5a9ac0e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.329s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.899983] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-127a8537-6efd-4177-aa79-53d082b3e795 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.909312] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39d4961d-595e-47e8-b14a-ee6a9f91dac9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.944093] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17bc9b49-1bf9-4729-849e-0789e6f666e2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.952274] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b5093c9-3264-49d2-ae8f-ab6c0698862e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.968111] env[61906]: DEBUG nova.compute.provider_tree [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 866.079221] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d7636e0-9fc9-4e8a-9de2-6ffeadf901f4 tempest-ServerAddressesNegativeTestJSON-1682752707 tempest-ServerAddressesNegativeTestJSON-1682752707-project-member] Lock "976021e6-0e38-4295-bdd4-83f00a7b7be0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.681s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.298508] env[61906]: DEBUG nova.compute.manager [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 866.472549] env[61906]: DEBUG nova.scheduler.client.report [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 866.586577] env[61906]: DEBUG nova.compute.manager [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 866.615740] env[61906]: DEBUG oslo_concurrency.lockutils [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Acquiring lock "7e05fee8-842f-4e03-853a-58bac5a9ac0e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.615740] env[61906]: DEBUG oslo_concurrency.lockutils [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Lock "7e05fee8-842f-4e03-853a-58bac5a9ac0e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.615740] env[61906]: DEBUG oslo_concurrency.lockutils [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Acquiring lock "7e05fee8-842f-4e03-853a-58bac5a9ac0e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.616038] env[61906]: DEBUG oslo_concurrency.lockutils [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Lock "7e05fee8-842f-4e03-853a-58bac5a9ac0e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.616197] env[61906]: DEBUG oslo_concurrency.lockutils [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Lock "7e05fee8-842f-4e03-853a-58bac5a9ac0e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.619209] env[61906]: INFO nova.compute.manager [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Terminating instance [ 866.621213] env[61906]: DEBUG nova.compute.manager [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 866.621776] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 866.622418] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62a8d590-e093-4a71-b6ec-a1620938474f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.632404] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 866.634141] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b882c6ff-4ed8-4b0f-84a5-2c6fa81473a8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.640906] env[61906]: DEBUG oslo_vmware.api [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Waiting for the task: (returnval){ [ 866.640906] env[61906]: value = "task-1333175" [ 866.640906] env[61906]: _type = "Task" [ 866.640906] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.651154] env[61906]: DEBUG nova.compute.manager [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 866.654108] env[61906]: DEBUG oslo_vmware.api [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Task: {'id': task-1333175, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.694034] env[61906]: DEBUG nova.virt.hardware [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 866.694034] env[61906]: DEBUG nova.virt.hardware [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 866.694034] env[61906]: DEBUG nova.virt.hardware [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 866.694034] env[61906]: DEBUG nova.virt.hardware [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 866.694304] env[61906]: DEBUG nova.virt.hardware [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 866.694466] env[61906]: DEBUG nova.virt.hardware [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 866.694748] env[61906]: DEBUG nova.virt.hardware [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 866.694944] env[61906]: DEBUG nova.virt.hardware [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 866.695200] env[61906]: DEBUG nova.virt.hardware [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 866.695439] env[61906]: DEBUG nova.virt.hardware [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 866.695736] env[61906]: DEBUG nova.virt.hardware [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 866.697311] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d2f5989-6ac5-4710-91ff-50a5c097fdfe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.709496] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c5e5568-36bc-4041-853c-cd8711daf83c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.827184] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.979069] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.357s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.982477] env[61906]: DEBUG nova.compute.manager [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 866.985452] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.287s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.987285] env[61906]: INFO nova.compute.claims [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 867.114727] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.152152] env[61906]: DEBUG oslo_vmware.api [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Task: {'id': task-1333175, 'name': PowerOffVM_Task, 'duration_secs': 0.199363} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.152499] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 867.152886] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 867.153208] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-894615cb-5ef1-48c4-ae86-7a21e3a4d367 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.257773] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 867.258023] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 867.258274] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Deleting the datastore file [datastore2] 7e05fee8-842f-4e03-853a-58bac5a9ac0e {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 867.258509] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8f8026e9-6c0c-4ab5-9061-e5f73a2fdbf7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.265956] env[61906]: DEBUG oslo_vmware.api [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Waiting for the task: (returnval){ [ 867.265956] env[61906]: value = "task-1333177" [ 867.265956] env[61906]: _type = "Task" [ 867.265956] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.280301] env[61906]: DEBUG oslo_vmware.api [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Task: {'id': task-1333177, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.492994] env[61906]: DEBUG nova.compute.utils [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 867.498108] env[61906]: DEBUG nova.compute.manager [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 867.498301] env[61906]: DEBUG nova.network.neutron [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 867.538102] env[61906]: DEBUG nova.compute.manager [req-ccd8a92e-7259-4802-b5ca-744f58712f14 req-2e99cc46-5809-4044-bd57-a680e214d18a service nova] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Received event network-vif-plugged-61a42623-e2e5-4245-9560-be57af9a8397 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.538370] env[61906]: DEBUG oslo_concurrency.lockutils [req-ccd8a92e-7259-4802-b5ca-744f58712f14 req-2e99cc46-5809-4044-bd57-a680e214d18a service nova] Acquiring lock "9f1253e6-4727-4dca-bad2-04f0c5424cdb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.538583] env[61906]: DEBUG oslo_concurrency.lockutils [req-ccd8a92e-7259-4802-b5ca-744f58712f14 req-2e99cc46-5809-4044-bd57-a680e214d18a service nova] Lock "9f1253e6-4727-4dca-bad2-04f0c5424cdb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.538786] env[61906]: DEBUG oslo_concurrency.lockutils [req-ccd8a92e-7259-4802-b5ca-744f58712f14 req-2e99cc46-5809-4044-bd57-a680e214d18a service nova] Lock "9f1253e6-4727-4dca-bad2-04f0c5424cdb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.538965] env[61906]: DEBUG nova.compute.manager [req-ccd8a92e-7259-4802-b5ca-744f58712f14 req-2e99cc46-5809-4044-bd57-a680e214d18a service nova] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] No waiting events found dispatching network-vif-plugged-61a42623-e2e5-4245-9560-be57af9a8397 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 867.539170] env[61906]: WARNING nova.compute.manager [req-ccd8a92e-7259-4802-b5ca-744f58712f14 req-2e99cc46-5809-4044-bd57-a680e214d18a service nova] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Received unexpected event network-vif-plugged-61a42623-e2e5-4245-9560-be57af9a8397 for instance with vm_state building and task_state spawning. [ 867.629114] env[61906]: DEBUG nova.policy [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f33ae649c28a45688c8662d4b3a29c9a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ac8ee7432c4a414d9d6d7ce5f60ced7a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 867.738037] env[61906]: DEBUG nova.network.neutron [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Successfully updated port: 61a42623-e2e5-4245-9560-be57af9a8397 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 867.781278] env[61906]: DEBUG oslo_vmware.api [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Task: {'id': task-1333177, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.2895} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.781552] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 867.782477] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 867.782477] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 867.782477] env[61906]: INFO nova.compute.manager [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Took 1.16 seconds to destroy the instance on the hypervisor. [ 867.782477] env[61906]: DEBUG oslo.service.loopingcall [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.782639] env[61906]: DEBUG nova.compute.manager [-] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 867.782639] env[61906]: DEBUG nova.network.neutron [-] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 868.000893] env[61906]: DEBUG nova.compute.manager [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 868.236735] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a360d26-a082-4dc2-8355-d359aefe2690 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.244123] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "refresh_cache-9f1253e6-4727-4dca-bad2-04f0c5424cdb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.245094] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired lock "refresh_cache-9f1253e6-4727-4dca-bad2-04f0c5424cdb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.245094] env[61906]: DEBUG nova.network.neutron [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 868.246763] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-938dc01c-da31-4201-a26c-9953030f967b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.284252] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bddaab4-dc44-4bb7-bdc4-1def686cafb9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.298168] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b43081d-2946-42a3-b3b1-843f6e21dee4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.303446] env[61906]: DEBUG nova.network.neutron [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Successfully created port: b6dc013b-172b-44b2-9f3c-a78cb9245964 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 868.315981] env[61906]: DEBUG nova.compute.provider_tree [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.790304] env[61906]: DEBUG nova.network.neutron [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 868.822530] env[61906]: DEBUG nova.scheduler.client.report [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 869.016226] env[61906]: DEBUG nova.compute.manager [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 869.048045] env[61906]: DEBUG nova.virt.hardware [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 869.048045] env[61906]: DEBUG nova.virt.hardware [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 869.048045] env[61906]: DEBUG nova.virt.hardware [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 869.048045] env[61906]: DEBUG nova.virt.hardware [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 869.048045] env[61906]: DEBUG nova.virt.hardware [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 869.048045] env[61906]: DEBUG nova.virt.hardware [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 869.048045] env[61906]: DEBUG nova.virt.hardware [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 869.048045] env[61906]: DEBUG nova.virt.hardware [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 869.048045] env[61906]: DEBUG nova.virt.hardware [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 869.048757] env[61906]: DEBUG nova.virt.hardware [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 869.049126] env[61906]: DEBUG nova.virt.hardware [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 869.050966] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78543795-e5e5-4b78-9f62-9d788c7e77e0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.059466] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f378177d-83d8-4022-b3fb-b1f263cbf728 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.196772] env[61906]: DEBUG nova.network.neutron [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Updating instance_info_cache with network_info: [{"id": "61a42623-e2e5-4245-9560-be57af9a8397", "address": "fa:16:3e:ff:cc:9a", "network": {"id": "af039bdd-ffd2-483d-9655-e7df48936290", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-111531540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f7fc2042e6549cabe98838990f35842", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61a42623-e2", "ovs_interfaceid": "61a42623-e2e5-4245-9560-be57af9a8397", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.335416] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.350s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.335712] env[61906]: DEBUG nova.compute.manager [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 869.338270] env[61906]: DEBUG oslo_concurrency.lockutils [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.390s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.340136] env[61906]: INFO nova.compute.claims [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 869.357474] env[61906]: DEBUG nova.network.neutron [-] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.572168] env[61906]: DEBUG nova.compute.manager [req-2028345f-5fee-44f0-a654-34bdf66e7071 req-ce1981e9-53c8-41bb-ace3-f78d995dfa7a service nova] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Received event network-changed-61a42623-e2e5-4245-9560-be57af9a8397 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.572379] env[61906]: DEBUG nova.compute.manager [req-2028345f-5fee-44f0-a654-34bdf66e7071 req-ce1981e9-53c8-41bb-ace3-f78d995dfa7a service nova] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Refreshing instance network info cache due to event network-changed-61a42623-e2e5-4245-9560-be57af9a8397. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 869.572576] env[61906]: DEBUG oslo_concurrency.lockutils [req-2028345f-5fee-44f0-a654-34bdf66e7071 req-ce1981e9-53c8-41bb-ace3-f78d995dfa7a service nova] Acquiring lock "refresh_cache-9f1253e6-4727-4dca-bad2-04f0c5424cdb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.702666] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Releasing lock "refresh_cache-9f1253e6-4727-4dca-bad2-04f0c5424cdb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.703013] env[61906]: DEBUG nova.compute.manager [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Instance network_info: |[{"id": "61a42623-e2e5-4245-9560-be57af9a8397", "address": "fa:16:3e:ff:cc:9a", "network": {"id": "af039bdd-ffd2-483d-9655-e7df48936290", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-111531540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f7fc2042e6549cabe98838990f35842", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61a42623-e2", "ovs_interfaceid": "61a42623-e2e5-4245-9560-be57af9a8397", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 869.703335] env[61906]: DEBUG oslo_concurrency.lockutils [req-2028345f-5fee-44f0-a654-34bdf66e7071 req-ce1981e9-53c8-41bb-ace3-f78d995dfa7a service nova] Acquired lock "refresh_cache-9f1253e6-4727-4dca-bad2-04f0c5424cdb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.703515] env[61906]: DEBUG nova.network.neutron [req-2028345f-5fee-44f0-a654-34bdf66e7071 req-ce1981e9-53c8-41bb-ace3-f78d995dfa7a service nova] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Refreshing network info cache for port 61a42623-e2e5-4245-9560-be57af9a8397 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 869.704830] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ff:cc:9a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '19440099-773e-4a31-b82e-84a4daa5d8fe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '61a42623-e2e5-4245-9560-be57af9a8397', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 869.714344] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Creating folder: Project (4f7fc2042e6549cabe98838990f35842). Parent ref: group-v284713. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 869.717475] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f9327612-1f70-405b-baa2-8c46b6f94239 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.729392] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Created folder: Project (4f7fc2042e6549cabe98838990f35842) in parent group-v284713. [ 869.729392] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Creating folder: Instances. Parent ref: group-v284741. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 869.729767] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5219a4a0-7f60-45b3-82ee-41ae410b642c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.738437] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Created folder: Instances in parent group-v284741. [ 869.738794] env[61906]: DEBUG oslo.service.loopingcall [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 869.738846] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 869.739048] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3bb4cdf8-5925-4392-9e8c-2d87e3385426 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.762987] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 869.762987] env[61906]: value = "task-1333180" [ 869.762987] env[61906]: _type = "Task" [ 869.762987] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.771025] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333180, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.845975] env[61906]: DEBUG nova.compute.utils [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 869.849389] env[61906]: DEBUG nova.compute.manager [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 869.849560] env[61906]: DEBUG nova.network.neutron [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 869.861206] env[61906]: INFO nova.compute.manager [-] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Took 2.08 seconds to deallocate network for instance. [ 869.900856] env[61906]: DEBUG nova.policy [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ff655cb32ff433fa79e528dce0626be', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '454208ab90fa4d029c222814aeaa0b49', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 870.028529] env[61906]: DEBUG nova.network.neutron [req-2028345f-5fee-44f0-a654-34bdf66e7071 req-ce1981e9-53c8-41bb-ace3-f78d995dfa7a service nova] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Updated VIF entry in instance network info cache for port 61a42623-e2e5-4245-9560-be57af9a8397. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 870.028935] env[61906]: DEBUG nova.network.neutron [req-2028345f-5fee-44f0-a654-34bdf66e7071 req-ce1981e9-53c8-41bb-ace3-f78d995dfa7a service nova] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Updating instance_info_cache with network_info: [{"id": "61a42623-e2e5-4245-9560-be57af9a8397", "address": "fa:16:3e:ff:cc:9a", "network": {"id": "af039bdd-ffd2-483d-9655-e7df48936290", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-111531540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f7fc2042e6549cabe98838990f35842", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61a42623-e2", "ovs_interfaceid": "61a42623-e2e5-4245-9560-be57af9a8397", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.273007] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333180, 'name': CreateVM_Task, 'duration_secs': 0.301025} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.273255] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 870.273967] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.274255] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.274726] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 870.275013] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c55a513-aaa6-4b28-a51e-597eabaf7c99 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.279668] env[61906]: DEBUG oslo_vmware.api [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 870.279668] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52b62cef-71e3-def8-b4ab-a5117e55aca0" [ 870.279668] env[61906]: _type = "Task" [ 870.279668] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.287489] env[61906]: DEBUG oslo_vmware.api [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52b62cef-71e3-def8-b4ab-a5117e55aca0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.355649] env[61906]: DEBUG nova.compute.manager [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 870.375433] env[61906]: DEBUG oslo_concurrency.lockutils [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.434483] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Acquiring lock "0e86bfdf-9002-4a2d-a2ac-af35921b020e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.434741] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Lock "0e86bfdf-9002-4a2d-a2ac-af35921b020e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.507440] env[61906]: DEBUG nova.network.neutron [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Successfully created port: 531cc05e-2f49-4b47-84da-bef5affc1bf2 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 870.532616] env[61906]: DEBUG oslo_concurrency.lockutils [req-2028345f-5fee-44f0-a654-34bdf66e7071 req-ce1981e9-53c8-41bb-ace3-f78d995dfa7a service nova] Releasing lock "refresh_cache-9f1253e6-4727-4dca-bad2-04f0c5424cdb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.532991] env[61906]: DEBUG nova.compute.manager [req-2028345f-5fee-44f0-a654-34bdf66e7071 req-ce1981e9-53c8-41bb-ace3-f78d995dfa7a service nova] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Received event network-vif-deleted-d5ae18fb-369d-4c5b-9efe-a7d2898521e9 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.533246] env[61906]: DEBUG nova.compute.manager [req-2028345f-5fee-44f0-a654-34bdf66e7071 req-ce1981e9-53c8-41bb-ace3-f78d995dfa7a service nova] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Received event network-vif-deleted-c156b7ea-028a-4b34-93cd-25d384ab5e76 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.632282] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a48348d-6158-4845-a2d7-01c3a157e9c5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.638512] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d63ede6-bb98-49bc-b37f-8a99860e6a28 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.672611] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79a25069-3de5-4cc5-8123-f9a729261972 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.682953] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9dd9edf-6082-4de5-9275-0979b052f6aa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.698628] env[61906]: DEBUG nova.compute.provider_tree [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 870.700898] env[61906]: DEBUG nova.network.neutron [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Successfully updated port: b6dc013b-172b-44b2-9f3c-a78cb9245964 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 870.791781] env[61906]: DEBUG oslo_vmware.api [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52b62cef-71e3-def8-b4ab-a5117e55aca0, 'name': SearchDatastore_Task, 'duration_secs': 0.00896} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.792185] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.792496] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 870.792829] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.793010] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.793266] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 870.793567] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-576e6d71-4a57-415f-9922-f46cd8ea8ffe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.801546] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 870.801728] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 870.802446] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b3a1e37-d77e-4366-9721-90e7f0416712 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.807584] env[61906]: DEBUG oslo_vmware.api [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 870.807584] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]524af98b-ff0c-8c74-1c9e-3a53c77baddd" [ 870.807584] env[61906]: _type = "Task" [ 870.807584] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.814943] env[61906]: DEBUG oslo_vmware.api [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]524af98b-ff0c-8c74-1c9e-3a53c77baddd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.203511] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Acquiring lock "refresh_cache-80a6a0b8-17c3-46e8-8861-6d1759a19111" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.203796] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Acquired lock "refresh_cache-80a6a0b8-17c3-46e8-8861-6d1759a19111" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.203796] env[61906]: DEBUG nova.network.neutron [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 871.205888] env[61906]: DEBUG nova.scheduler.client.report [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 871.318676] env[61906]: DEBUG oslo_vmware.api [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]524af98b-ff0c-8c74-1c9e-3a53c77baddd, 'name': SearchDatastore_Task, 'duration_secs': 0.008315} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.319482] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6aa04e3d-4296-466d-9fc7-51aabb3d5c83 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.325084] env[61906]: DEBUG oslo_vmware.api [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 871.325084] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]521ff660-60a5-9e35-1b7d-9045ebb2bf24" [ 871.325084] env[61906]: _type = "Task" [ 871.325084] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.332867] env[61906]: DEBUG oslo_vmware.api [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]521ff660-60a5-9e35-1b7d-9045ebb2bf24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.369534] env[61906]: DEBUG nova.compute.manager [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 871.394370] env[61906]: DEBUG nova.virt.hardware [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 871.394620] env[61906]: DEBUG nova.virt.hardware [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 871.394775] env[61906]: DEBUG nova.virt.hardware [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 871.394954] env[61906]: DEBUG nova.virt.hardware [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 871.395112] env[61906]: DEBUG nova.virt.hardware [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 871.395259] env[61906]: DEBUG nova.virt.hardware [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 871.395487] env[61906]: DEBUG nova.virt.hardware [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 871.395664] env[61906]: DEBUG nova.virt.hardware [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 871.395831] env[61906]: DEBUG nova.virt.hardware [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 871.395990] env[61906]: DEBUG nova.virt.hardware [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 871.396226] env[61906]: DEBUG nova.virt.hardware [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 871.397007] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5249797-225e-4550-b766-aab7b786f8db {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.404853] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a5b427d-74f4-4cab-8775-bdf46d3937b2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.607913] env[61906]: DEBUG nova.compute.manager [req-68497e17-111f-40f7-ba94-0a510f5b1f56 req-c72448af-6df8-4ebd-a99f-2f263195b9fc service nova] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Received event network-vif-plugged-b6dc013b-172b-44b2-9f3c-a78cb9245964 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.608160] env[61906]: DEBUG oslo_concurrency.lockutils [req-68497e17-111f-40f7-ba94-0a510f5b1f56 req-c72448af-6df8-4ebd-a99f-2f263195b9fc service nova] Acquiring lock "80a6a0b8-17c3-46e8-8861-6d1759a19111-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.608370] env[61906]: DEBUG oslo_concurrency.lockutils [req-68497e17-111f-40f7-ba94-0a510f5b1f56 req-c72448af-6df8-4ebd-a99f-2f263195b9fc service nova] Lock "80a6a0b8-17c3-46e8-8861-6d1759a19111-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.608539] env[61906]: DEBUG oslo_concurrency.lockutils [req-68497e17-111f-40f7-ba94-0a510f5b1f56 req-c72448af-6df8-4ebd-a99f-2f263195b9fc service nova] Lock "80a6a0b8-17c3-46e8-8861-6d1759a19111-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.608703] env[61906]: DEBUG nova.compute.manager [req-68497e17-111f-40f7-ba94-0a510f5b1f56 req-c72448af-6df8-4ebd-a99f-2f263195b9fc service nova] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] No waiting events found dispatching network-vif-plugged-b6dc013b-172b-44b2-9f3c-a78cb9245964 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 871.608864] env[61906]: WARNING nova.compute.manager [req-68497e17-111f-40f7-ba94-0a510f5b1f56 req-c72448af-6df8-4ebd-a99f-2f263195b9fc service nova] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Received unexpected event network-vif-plugged-b6dc013b-172b-44b2-9f3c-a78cb9245964 for instance with vm_state building and task_state spawning. [ 871.609031] env[61906]: DEBUG nova.compute.manager [req-68497e17-111f-40f7-ba94-0a510f5b1f56 req-c72448af-6df8-4ebd-a99f-2f263195b9fc service nova] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Received event network-changed-b6dc013b-172b-44b2-9f3c-a78cb9245964 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.609190] env[61906]: DEBUG nova.compute.manager [req-68497e17-111f-40f7-ba94-0a510f5b1f56 req-c72448af-6df8-4ebd-a99f-2f263195b9fc service nova] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Refreshing instance network info cache due to event network-changed-b6dc013b-172b-44b2-9f3c-a78cb9245964. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 871.609353] env[61906]: DEBUG oslo_concurrency.lockutils [req-68497e17-111f-40f7-ba94-0a510f5b1f56 req-c72448af-6df8-4ebd-a99f-2f263195b9fc service nova] Acquiring lock "refresh_cache-80a6a0b8-17c3-46e8-8861-6d1759a19111" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.712520] env[61906]: DEBUG oslo_concurrency.lockutils [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.374s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.713124] env[61906]: DEBUG nova.compute.manager [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 871.715749] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.039s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.717123] env[61906]: INFO nova.compute.claims [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 871.768013] env[61906]: DEBUG nova.network.neutron [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 871.835057] env[61906]: DEBUG oslo_vmware.api [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]521ff660-60a5-9e35-1b7d-9045ebb2bf24, 'name': SearchDatastore_Task, 'duration_secs': 0.008237} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.835340] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.835599] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 9f1253e6-4727-4dca-bad2-04f0c5424cdb/9f1253e6-4727-4dca-bad2-04f0c5424cdb.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 871.835854] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fb6e1c74-42ba-4346-8095-2c7bd931a34e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.842325] env[61906]: DEBUG oslo_vmware.api [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 871.842325] env[61906]: value = "task-1333181" [ 871.842325] env[61906]: _type = "Task" [ 871.842325] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.850155] env[61906]: DEBUG oslo_vmware.api [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333181, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.991487] env[61906]: DEBUG nova.network.neutron [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Updating instance_info_cache with network_info: [{"id": "b6dc013b-172b-44b2-9f3c-a78cb9245964", "address": "fa:16:3e:09:3c:46", "network": {"id": "e9e4d889-5e38-4828-ba70-3ac52f8b72fa", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-389623216-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ac8ee7432c4a414d9d6d7ce5f60ced7a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "822050c7-1845-485d-b87e-73778d21c33c", "external-id": "nsx-vlan-transportzone-701", "segmentation_id": 701, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6dc013b-17", "ovs_interfaceid": "b6dc013b-172b-44b2-9f3c-a78cb9245964", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.222039] env[61906]: DEBUG nova.compute.utils [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 872.227321] env[61906]: DEBUG nova.compute.manager [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 872.227321] env[61906]: DEBUG nova.network.neutron [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 872.321592] env[61906]: DEBUG nova.policy [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0de9abfd085f4a43a878da8448615bc4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e36ff7a500ef444cbf5a168c5a48208b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 872.341988] env[61906]: DEBUG nova.network.neutron [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Successfully updated port: 531cc05e-2f49-4b47-84da-bef5affc1bf2 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 872.354166] env[61906]: DEBUG oslo_vmware.api [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333181, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475584} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.354975] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 9f1253e6-4727-4dca-bad2-04f0c5424cdb/9f1253e6-4727-4dca-bad2-04f0c5424cdb.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 872.355212] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 872.355465] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d33bacec-2a72-4e9c-88d6-1c873bc6b4ff {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.361908] env[61906]: DEBUG oslo_vmware.api [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 872.361908] env[61906]: value = "task-1333182" [ 872.361908] env[61906]: _type = "Task" [ 872.361908] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.369606] env[61906]: DEBUG oslo_vmware.api [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333182, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.496503] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Releasing lock "refresh_cache-80a6a0b8-17c3-46e8-8861-6d1759a19111" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.496813] env[61906]: DEBUG nova.compute.manager [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Instance network_info: |[{"id": "b6dc013b-172b-44b2-9f3c-a78cb9245964", "address": "fa:16:3e:09:3c:46", "network": {"id": "e9e4d889-5e38-4828-ba70-3ac52f8b72fa", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-389623216-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ac8ee7432c4a414d9d6d7ce5f60ced7a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "822050c7-1845-485d-b87e-73778d21c33c", "external-id": "nsx-vlan-transportzone-701", "segmentation_id": 701, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6dc013b-17", "ovs_interfaceid": "b6dc013b-172b-44b2-9f3c-a78cb9245964", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 872.497139] env[61906]: DEBUG oslo_concurrency.lockutils [req-68497e17-111f-40f7-ba94-0a510f5b1f56 req-c72448af-6df8-4ebd-a99f-2f263195b9fc service nova] Acquired lock "refresh_cache-80a6a0b8-17c3-46e8-8861-6d1759a19111" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.497324] env[61906]: DEBUG nova.network.neutron [req-68497e17-111f-40f7-ba94-0a510f5b1f56 req-c72448af-6df8-4ebd-a99f-2f263195b9fc service nova] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Refreshing network info cache for port b6dc013b-172b-44b2-9f3c-a78cb9245964 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 872.498488] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:09:3c:46', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '822050c7-1845-485d-b87e-73778d21c33c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b6dc013b-172b-44b2-9f3c-a78cb9245964', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 872.505717] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Creating folder: Project (ac8ee7432c4a414d9d6d7ce5f60ced7a). Parent ref: group-v284713. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 872.506660] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7671e348-7a87-402b-a076-5b4d4b54e855 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.518225] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Created folder: Project (ac8ee7432c4a414d9d6d7ce5f60ced7a) in parent group-v284713. [ 872.518404] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Creating folder: Instances. Parent ref: group-v284744. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 872.518627] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8c4ef494-7fc8-4fce-87a5-ed3e9df9628e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.527035] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Created folder: Instances in parent group-v284744. [ 872.527872] env[61906]: DEBUG oslo.service.loopingcall [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.527872] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 872.527872] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-027ab24e-6ea4-4573-b2b7-ed248972d6c0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.545693] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 872.545693] env[61906]: value = "task-1333185" [ 872.545693] env[61906]: _type = "Task" [ 872.545693] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.553094] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333185, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.593734] env[61906]: DEBUG nova.network.neutron [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Successfully created port: 3e04c5c7-8d7a-46ae-a2d7-52dc429650f9 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 872.726861] env[61906]: DEBUG nova.compute.manager [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 872.849516] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Acquiring lock "refresh_cache-d60575a6-af13-4d81-95d1-081d0f3e2ce6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.849636] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Acquired lock "refresh_cache-d60575a6-af13-4d81-95d1-081d0f3e2ce6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.849750] env[61906]: DEBUG nova.network.neutron [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 872.876878] env[61906]: DEBUG oslo_vmware.api [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333182, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065022} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.877171] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 872.877954] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-661ecbb5-99a7-4fc1-bcea-036ce56bb80b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.901083] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 9f1253e6-4727-4dca-bad2-04f0c5424cdb/9f1253e6-4727-4dca-bad2-04f0c5424cdb.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 872.903956] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ab22addc-2e14-4c6a-aac1-ca9410191b64 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.922978] env[61906]: DEBUG oslo_vmware.api [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 872.922978] env[61906]: value = "task-1333186" [ 872.922978] env[61906]: _type = "Task" [ 872.922978] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.931597] env[61906]: DEBUG oslo_vmware.api [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333186, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.962668] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5475a00-69ad-4eac-8b3e-d0cdb68b7b6b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.969879] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d7d3a41-89ec-482f-8507-7651b3c785ad {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.000651] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e7cc0a7-0637-43ed-83a0-52b94c7bfbc2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.011081] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e79dba7e-3555-4a8a-a62f-9fa365bba260 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.025025] env[61906]: DEBUG nova.compute.provider_tree [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.055578] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333185, 'name': CreateVM_Task, 'duration_secs': 0.299323} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.055746] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 873.056429] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.056593] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.056917] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 873.057183] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ba5bc7a-d486-462d-8b4c-429834c39133 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.063449] env[61906]: DEBUG oslo_vmware.api [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Waiting for the task: (returnval){ [ 873.063449] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]522f97d4-2132-d9d8-fd43-9fc028add1fc" [ 873.063449] env[61906]: _type = "Task" [ 873.063449] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.071164] env[61906]: DEBUG oslo_vmware.api [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]522f97d4-2132-d9d8-fd43-9fc028add1fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.291228] env[61906]: DEBUG nova.network.neutron [req-68497e17-111f-40f7-ba94-0a510f5b1f56 req-c72448af-6df8-4ebd-a99f-2f263195b9fc service nova] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Updated VIF entry in instance network info cache for port b6dc013b-172b-44b2-9f3c-a78cb9245964. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 873.291711] env[61906]: DEBUG nova.network.neutron [req-68497e17-111f-40f7-ba94-0a510f5b1f56 req-c72448af-6df8-4ebd-a99f-2f263195b9fc service nova] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Updating instance_info_cache with network_info: [{"id": "b6dc013b-172b-44b2-9f3c-a78cb9245964", "address": "fa:16:3e:09:3c:46", "network": {"id": "e9e4d889-5e38-4828-ba70-3ac52f8b72fa", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-389623216-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ac8ee7432c4a414d9d6d7ce5f60ced7a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "822050c7-1845-485d-b87e-73778d21c33c", "external-id": "nsx-vlan-transportzone-701", "segmentation_id": 701, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6dc013b-17", "ovs_interfaceid": "b6dc013b-172b-44b2-9f3c-a78cb9245964", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.384904] env[61906]: DEBUG nova.network.neutron [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 873.435368] env[61906]: DEBUG oslo_vmware.api [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333186, 'name': ReconfigVM_Task, 'duration_secs': 0.273115} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.435652] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 9f1253e6-4727-4dca-bad2-04f0c5424cdb/9f1253e6-4727-4dca-bad2-04f0c5424cdb.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 873.436276] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e75b53e2-4d41-4260-9eb6-26806392bb2a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.442819] env[61906]: DEBUG oslo_vmware.api [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 873.442819] env[61906]: value = "task-1333187" [ 873.442819] env[61906]: _type = "Task" [ 873.442819] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.450576] env[61906]: DEBUG oslo_vmware.api [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333187, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.522263] env[61906]: DEBUG nova.network.neutron [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Updating instance_info_cache with network_info: [{"id": "531cc05e-2f49-4b47-84da-bef5affc1bf2", "address": "fa:16:3e:61:9c:75", "network": {"id": "9dd01f29-bcc2-4dec-8386-0e4967f7f1d1", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1594578243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "454208ab90fa4d029c222814aeaa0b49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2e45023-22b5-458b-826e-9b7eb69ba028", "external-id": "nsx-vlan-transportzone-614", "segmentation_id": 614, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap531cc05e-2f", "ovs_interfaceid": "531cc05e-2f49-4b47-84da-bef5affc1bf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.527728] env[61906]: DEBUG nova.scheduler.client.report [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 873.574529] env[61906]: DEBUG oslo_vmware.api [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]522f97d4-2132-d9d8-fd43-9fc028add1fc, 'name': SearchDatastore_Task, 'duration_secs': 0.009267} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.575112] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.575354] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 873.575598] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.575732] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.575908] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 873.576180] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-23ad06be-5a38-44a0-8100-672099b99056 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.584258] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 873.584421] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 873.585091] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37485a80-ae7a-4e2f-8848-1a9ffdab6b42 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.590041] env[61906]: DEBUG oslo_vmware.api [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Waiting for the task: (returnval){ [ 873.590041] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]522929f6-f715-fc8f-3150-0cba6394defa" [ 873.590041] env[61906]: _type = "Task" [ 873.590041] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.597042] env[61906]: DEBUG oslo_vmware.api [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]522929f6-f715-fc8f-3150-0cba6394defa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.632707] env[61906]: DEBUG nova.compute.manager [req-6e36c6f0-0100-4e79-9a17-ecb829718938 req-3213da84-b267-468f-b370-6b579534f97c service nova] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Received event network-vif-plugged-531cc05e-2f49-4b47-84da-bef5affc1bf2 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.632947] env[61906]: DEBUG oslo_concurrency.lockutils [req-6e36c6f0-0100-4e79-9a17-ecb829718938 req-3213da84-b267-468f-b370-6b579534f97c service nova] Acquiring lock "d60575a6-af13-4d81-95d1-081d0f3e2ce6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.633193] env[61906]: DEBUG oslo_concurrency.lockutils [req-6e36c6f0-0100-4e79-9a17-ecb829718938 req-3213da84-b267-468f-b370-6b579534f97c service nova] Lock "d60575a6-af13-4d81-95d1-081d0f3e2ce6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.633352] env[61906]: DEBUG oslo_concurrency.lockutils [req-6e36c6f0-0100-4e79-9a17-ecb829718938 req-3213da84-b267-468f-b370-6b579534f97c service nova] Lock "d60575a6-af13-4d81-95d1-081d0f3e2ce6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.633520] env[61906]: DEBUG nova.compute.manager [req-6e36c6f0-0100-4e79-9a17-ecb829718938 req-3213da84-b267-468f-b370-6b579534f97c service nova] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] No waiting events found dispatching network-vif-plugged-531cc05e-2f49-4b47-84da-bef5affc1bf2 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 873.633681] env[61906]: WARNING nova.compute.manager [req-6e36c6f0-0100-4e79-9a17-ecb829718938 req-3213da84-b267-468f-b370-6b579534f97c service nova] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Received unexpected event network-vif-plugged-531cc05e-2f49-4b47-84da-bef5affc1bf2 for instance with vm_state building and task_state spawning. [ 873.633837] env[61906]: DEBUG nova.compute.manager [req-6e36c6f0-0100-4e79-9a17-ecb829718938 req-3213da84-b267-468f-b370-6b579534f97c service nova] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Received event network-changed-531cc05e-2f49-4b47-84da-bef5affc1bf2 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.633985] env[61906]: DEBUG nova.compute.manager [req-6e36c6f0-0100-4e79-9a17-ecb829718938 req-3213da84-b267-468f-b370-6b579534f97c service nova] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Refreshing instance network info cache due to event network-changed-531cc05e-2f49-4b47-84da-bef5affc1bf2. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 873.634165] env[61906]: DEBUG oslo_concurrency.lockutils [req-6e36c6f0-0100-4e79-9a17-ecb829718938 req-3213da84-b267-468f-b370-6b579534f97c service nova] Acquiring lock "refresh_cache-d60575a6-af13-4d81-95d1-081d0f3e2ce6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.744014] env[61906]: DEBUG nova.compute.manager [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 873.768759] env[61906]: DEBUG nova.virt.hardware [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 873.769009] env[61906]: DEBUG nova.virt.hardware [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 873.769181] env[61906]: DEBUG nova.virt.hardware [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 873.769365] env[61906]: DEBUG nova.virt.hardware [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 873.769512] env[61906]: DEBUG nova.virt.hardware [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 873.769659] env[61906]: DEBUG nova.virt.hardware [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 873.769864] env[61906]: DEBUG nova.virt.hardware [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 873.770031] env[61906]: DEBUG nova.virt.hardware [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 873.770203] env[61906]: DEBUG nova.virt.hardware [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 873.770364] env[61906]: DEBUG nova.virt.hardware [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 873.770533] env[61906]: DEBUG nova.virt.hardware [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 873.771424] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00077007-5ed0-41df-881c-60fe402c9b81 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.779551] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f09d14a-d1e6-4237-a083-530e49211280 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.794851] env[61906]: DEBUG oslo_concurrency.lockutils [req-68497e17-111f-40f7-ba94-0a510f5b1f56 req-c72448af-6df8-4ebd-a99f-2f263195b9fc service nova] Releasing lock "refresh_cache-80a6a0b8-17c3-46e8-8861-6d1759a19111" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.952755] env[61906]: DEBUG oslo_vmware.api [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333187, 'name': Rename_Task, 'duration_secs': 0.138518} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.953079] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 873.953327] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-58d501c7-de30-47fa-ba95-20281c3105eb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.959573] env[61906]: DEBUG oslo_vmware.api [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 873.959573] env[61906]: value = "task-1333188" [ 873.959573] env[61906]: _type = "Task" [ 873.959573] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.966762] env[61906]: DEBUG oslo_vmware.api [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333188, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.024640] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Releasing lock "refresh_cache-d60575a6-af13-4d81-95d1-081d0f3e2ce6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.025072] env[61906]: DEBUG nova.compute.manager [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Instance network_info: |[{"id": "531cc05e-2f49-4b47-84da-bef5affc1bf2", "address": "fa:16:3e:61:9c:75", "network": {"id": "9dd01f29-bcc2-4dec-8386-0e4967f7f1d1", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1594578243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "454208ab90fa4d029c222814aeaa0b49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2e45023-22b5-458b-826e-9b7eb69ba028", "external-id": "nsx-vlan-transportzone-614", "segmentation_id": 614, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap531cc05e-2f", "ovs_interfaceid": "531cc05e-2f49-4b47-84da-bef5affc1bf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 874.025429] env[61906]: DEBUG oslo_concurrency.lockutils [req-6e36c6f0-0100-4e79-9a17-ecb829718938 req-3213da84-b267-468f-b370-6b579534f97c service nova] Acquired lock "refresh_cache-d60575a6-af13-4d81-95d1-081d0f3e2ce6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.025625] env[61906]: DEBUG nova.network.neutron [req-6e36c6f0-0100-4e79-9a17-ecb829718938 req-3213da84-b267-468f-b370-6b579534f97c service nova] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Refreshing network info cache for port 531cc05e-2f49-4b47-84da-bef5affc1bf2 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 874.027971] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:61:9c:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f2e45023-22b5-458b-826e-9b7eb69ba028', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '531cc05e-2f49-4b47-84da-bef5affc1bf2', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 874.035649] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Creating folder: Project (454208ab90fa4d029c222814aeaa0b49). Parent ref: group-v284713. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 874.036665] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.321s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.036952] env[61906]: DEBUG nova.compute.manager [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 874.039265] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6cbf1b02-d777-4f84-bd18-a8675269d496 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.041571] env[61906]: DEBUG oslo_concurrency.lockutils [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.479s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.043173] env[61906]: INFO nova.compute.claims [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 874.055693] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Created folder: Project (454208ab90fa4d029c222814aeaa0b49) in parent group-v284713. [ 874.055931] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Creating folder: Instances. Parent ref: group-v284747. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 874.056220] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a4fc25ce-57e2-4abb-9f67-29f15d882331 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.066370] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Created folder: Instances in parent group-v284747. [ 874.066749] env[61906]: DEBUG oslo.service.loopingcall [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 874.066927] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 874.067144] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-90c1aebc-65ce-4a7c-904c-0d4e2e64d97d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.089561] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 874.089561] env[61906]: value = "task-1333191" [ 874.089561] env[61906]: _type = "Task" [ 874.089561] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.103024] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333191, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.106636] env[61906]: DEBUG oslo_vmware.api [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]522929f6-f715-fc8f-3150-0cba6394defa, 'name': SearchDatastore_Task, 'duration_secs': 0.008542} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.107112] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f9f721d-873d-48be-a1c8-88a9cd629c5e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.110456] env[61906]: DEBUG nova.network.neutron [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Successfully updated port: 3e04c5c7-8d7a-46ae-a2d7-52dc429650f9 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 874.115234] env[61906]: DEBUG oslo_vmware.api [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Waiting for the task: (returnval){ [ 874.115234] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]523fb74a-448d-7460-b6f4-d17ab863c5e0" [ 874.115234] env[61906]: _type = "Task" [ 874.115234] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.125888] env[61906]: DEBUG oslo_vmware.api [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]523fb74a-448d-7460-b6f4-d17ab863c5e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.469652] env[61906]: DEBUG oslo_vmware.api [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333188, 'name': PowerOnVM_Task, 'duration_secs': 0.422188} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.470088] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 874.470224] env[61906]: INFO nova.compute.manager [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Took 7.82 seconds to spawn the instance on the hypervisor. [ 874.470405] env[61906]: DEBUG nova.compute.manager [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 874.471167] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9371a5c0-5b34-4d45-bfdd-ea7e61a9dff4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.543364] env[61906]: DEBUG nova.compute.utils [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 874.544714] env[61906]: DEBUG nova.compute.manager [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Not allocating networking since 'none' was specified. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 874.600763] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333191, 'name': CreateVM_Task, 'duration_secs': 0.311611} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.600928] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 874.601550] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.601710] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.602094] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 874.602347] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-040b9da5-49f1-49e0-9322-ac7b17e0f788 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.606372] env[61906]: DEBUG oslo_vmware.api [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Waiting for the task: (returnval){ [ 874.606372] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52e87ca1-079d-3495-f19d-cb004e1bdbf2" [ 874.606372] env[61906]: _type = "Task" [ 874.606372] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.613687] env[61906]: DEBUG oslo_concurrency.lockutils [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "refresh_cache-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.613822] env[61906]: DEBUG oslo_concurrency.lockutils [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired lock "refresh_cache-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.613968] env[61906]: DEBUG nova.network.neutron [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 874.614946] env[61906]: DEBUG oslo_vmware.api [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52e87ca1-079d-3495-f19d-cb004e1bdbf2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.625072] env[61906]: DEBUG oslo_vmware.api [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]523fb74a-448d-7460-b6f4-d17ab863c5e0, 'name': SearchDatastore_Task, 'duration_secs': 0.010246} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.625311] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.625557] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 80a6a0b8-17c3-46e8-8861-6d1759a19111/80a6a0b8-17c3-46e8-8861-6d1759a19111.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 874.625789] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2bcbefa4-8588-4a77-bf7b-080278e35ecd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.632615] env[61906]: DEBUG oslo_vmware.api [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Waiting for the task: (returnval){ [ 874.632615] env[61906]: value = "task-1333192" [ 874.632615] env[61906]: _type = "Task" [ 874.632615] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.639833] env[61906]: DEBUG oslo_vmware.api [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Task: {'id': task-1333192, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.736872] env[61906]: DEBUG nova.network.neutron [req-6e36c6f0-0100-4e79-9a17-ecb829718938 req-3213da84-b267-468f-b370-6b579534f97c service nova] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Updated VIF entry in instance network info cache for port 531cc05e-2f49-4b47-84da-bef5affc1bf2. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 874.737286] env[61906]: DEBUG nova.network.neutron [req-6e36c6f0-0100-4e79-9a17-ecb829718938 req-3213da84-b267-468f-b370-6b579534f97c service nova] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Updating instance_info_cache with network_info: [{"id": "531cc05e-2f49-4b47-84da-bef5affc1bf2", "address": "fa:16:3e:61:9c:75", "network": {"id": "9dd01f29-bcc2-4dec-8386-0e4967f7f1d1", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1594578243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "454208ab90fa4d029c222814aeaa0b49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2e45023-22b5-458b-826e-9b7eb69ba028", "external-id": "nsx-vlan-transportzone-614", "segmentation_id": 614, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap531cc05e-2f", "ovs_interfaceid": "531cc05e-2f49-4b47-84da-bef5affc1bf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.987048] env[61906]: INFO nova.compute.manager [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Took 34.82 seconds to build instance. [ 875.046768] env[61906]: DEBUG nova.compute.manager [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 875.120826] env[61906]: DEBUG oslo_vmware.api [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52e87ca1-079d-3495-f19d-cb004e1bdbf2, 'name': SearchDatastore_Task, 'duration_secs': 0.00903} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.121205] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.121445] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 875.121677] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.121823] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.122032] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 875.123304] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-70cfa3ef-52d5-4644-bfb3-8fdfb806a9b3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.130464] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 875.130671] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 875.131384] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3813378d-2bd9-4ed2-b5ea-f58a6cd32d3b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.144294] env[61906]: DEBUG oslo_vmware.api [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Task: {'id': task-1333192, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465174} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.144983] env[61906]: DEBUG oslo_vmware.api [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Waiting for the task: (returnval){ [ 875.144983] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5295a6cd-e393-0dae-53f7-081e0a29dd7b" [ 875.144983] env[61906]: _type = "Task" [ 875.144983] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.144983] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 80a6a0b8-17c3-46e8-8861-6d1759a19111/80a6a0b8-17c3-46e8-8861-6d1759a19111.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 875.144983] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 875.145273] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b7553ac0-a276-4452-bcc7-32efdc297a9e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.152290] env[61906]: DEBUG nova.network.neutron [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 875.160673] env[61906]: DEBUG oslo_vmware.api [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5295a6cd-e393-0dae-53f7-081e0a29dd7b, 'name': SearchDatastore_Task, 'duration_secs': 0.008183} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.162054] env[61906]: DEBUG oslo_vmware.api [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Waiting for the task: (returnval){ [ 875.162054] env[61906]: value = "task-1333193" [ 875.162054] env[61906]: _type = "Task" [ 875.162054] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.162259] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e5aa5fb-5e56-40bd-8c34-4864c7df36b5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.174049] env[61906]: DEBUG oslo_vmware.api [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Waiting for the task: (returnval){ [ 875.174049] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52b80731-000b-ac1c-181d-c8b0db7c4b43" [ 875.174049] env[61906]: _type = "Task" [ 875.174049] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.178851] env[61906]: DEBUG oslo_vmware.api [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Task: {'id': task-1333193, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.189588] env[61906]: DEBUG oslo_vmware.api [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52b80731-000b-ac1c-181d-c8b0db7c4b43, 'name': SearchDatastore_Task, 'duration_secs': 0.008514} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.189846] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.190152] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] d60575a6-af13-4d81-95d1-081d0f3e2ce6/d60575a6-af13-4d81-95d1-081d0f3e2ce6.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 875.190428] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cea38563-3ff0-4135-addc-7d94b511522a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.197539] env[61906]: DEBUG oslo_vmware.api [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Waiting for the task: (returnval){ [ 875.197539] env[61906]: value = "task-1333194" [ 875.197539] env[61906]: _type = "Task" [ 875.197539] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.205398] env[61906]: DEBUG oslo_vmware.api [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333194, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.242174] env[61906]: DEBUG oslo_concurrency.lockutils [req-6e36c6f0-0100-4e79-9a17-ecb829718938 req-3213da84-b267-468f-b370-6b579534f97c service nova] Releasing lock "refresh_cache-d60575a6-af13-4d81-95d1-081d0f3e2ce6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.255403] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc1aaa3a-3a01-4553-afc4-247285bb4fe2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.262361] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1d9068f-7afa-4efd-99e2-cdfb5e1eb7e9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.294832] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba80064-32f0-4e9b-b6e6-3c62599edc8d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.302195] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efcc44af-f7d9-4e6e-9335-296cc120e693 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.316737] env[61906]: DEBUG nova.compute.provider_tree [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 875.331282] env[61906]: DEBUG nova.network.neutron [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Updating instance_info_cache with network_info: [{"id": "3e04c5c7-8d7a-46ae-a2d7-52dc429650f9", "address": "fa:16:3e:f1:15:76", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e04c5c7-8d", "ovs_interfaceid": "3e04c5c7-8d7a-46ae-a2d7-52dc429650f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.490773] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6998fb0-4054-456c-aa76-9869f0b97351 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "9f1253e6-4727-4dca-bad2-04f0c5424cdb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.846s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.677037] env[61906]: DEBUG oslo_vmware.api [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Task: {'id': task-1333193, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063481} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.677350] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 875.678438] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39478185-97a1-4849-8ecb-4698b9aa7641 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.710819] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 80a6a0b8-17c3-46e8-8861-6d1759a19111/80a6a0b8-17c3-46e8-8861-6d1759a19111.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 875.711238] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d44bd0aa-72b6-4821-ac23-2fc08c87d071 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.737643] env[61906]: DEBUG nova.compute.manager [req-22b8b5d4-2f1d-47f8-96f0-c65939176db0 req-9b219914-f1e7-4217-b7a3-2715c44c1381 service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Received event network-vif-plugged-3e04c5c7-8d7a-46ae-a2d7-52dc429650f9 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.737872] env[61906]: DEBUG oslo_concurrency.lockutils [req-22b8b5d4-2f1d-47f8-96f0-c65939176db0 req-9b219914-f1e7-4217-b7a3-2715c44c1381 service nova] Acquiring lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.738094] env[61906]: DEBUG oslo_concurrency.lockutils [req-22b8b5d4-2f1d-47f8-96f0-c65939176db0 req-9b219914-f1e7-4217-b7a3-2715c44c1381 service nova] Lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.738263] env[61906]: DEBUG oslo_concurrency.lockutils [req-22b8b5d4-2f1d-47f8-96f0-c65939176db0 req-9b219914-f1e7-4217-b7a3-2715c44c1381 service nova] Lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.738474] env[61906]: DEBUG nova.compute.manager [req-22b8b5d4-2f1d-47f8-96f0-c65939176db0 req-9b219914-f1e7-4217-b7a3-2715c44c1381 service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] No waiting events found dispatching network-vif-plugged-3e04c5c7-8d7a-46ae-a2d7-52dc429650f9 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 875.738593] env[61906]: WARNING nova.compute.manager [req-22b8b5d4-2f1d-47f8-96f0-c65939176db0 req-9b219914-f1e7-4217-b7a3-2715c44c1381 service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Received unexpected event network-vif-plugged-3e04c5c7-8d7a-46ae-a2d7-52dc429650f9 for instance with vm_state building and task_state spawning. [ 875.738747] env[61906]: DEBUG nova.compute.manager [req-22b8b5d4-2f1d-47f8-96f0-c65939176db0 req-9b219914-f1e7-4217-b7a3-2715c44c1381 service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Received event network-changed-3e04c5c7-8d7a-46ae-a2d7-52dc429650f9 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.738901] env[61906]: DEBUG nova.compute.manager [req-22b8b5d4-2f1d-47f8-96f0-c65939176db0 req-9b219914-f1e7-4217-b7a3-2715c44c1381 service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Refreshing instance network info cache due to event network-changed-3e04c5c7-8d7a-46ae-a2d7-52dc429650f9. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 875.739085] env[61906]: DEBUG oslo_concurrency.lockutils [req-22b8b5d4-2f1d-47f8-96f0-c65939176db0 req-9b219914-f1e7-4217-b7a3-2715c44c1381 service nova] Acquiring lock "refresh_cache-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.744583] env[61906]: DEBUG oslo_vmware.api [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Waiting for the task: (returnval){ [ 875.744583] env[61906]: value = "task-1333195" [ 875.744583] env[61906]: _type = "Task" [ 875.744583] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.748270] env[61906]: DEBUG oslo_vmware.api [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333194, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.522805} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.751827] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] d60575a6-af13-4d81-95d1-081d0f3e2ce6/d60575a6-af13-4d81-95d1-081d0f3e2ce6.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 875.752110] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 875.752450] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c2b258f5-dc24-429b-96f8-cbc5e35d0afe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.760507] env[61906]: DEBUG oslo_vmware.api [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Task: {'id': task-1333195, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.761866] env[61906]: DEBUG oslo_vmware.api [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Waiting for the task: (returnval){ [ 875.761866] env[61906]: value = "task-1333196" [ 875.761866] env[61906]: _type = "Task" [ 875.761866] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.771428] env[61906]: DEBUG oslo_vmware.api [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333196, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.819735] env[61906]: DEBUG nova.scheduler.client.report [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 875.834169] env[61906]: DEBUG oslo_concurrency.lockutils [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Releasing lock "refresh_cache-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.834513] env[61906]: DEBUG nova.compute.manager [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Instance network_info: |[{"id": "3e04c5c7-8d7a-46ae-a2d7-52dc429650f9", "address": "fa:16:3e:f1:15:76", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e04c5c7-8d", "ovs_interfaceid": "3e04c5c7-8d7a-46ae-a2d7-52dc429650f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 875.834803] env[61906]: DEBUG oslo_concurrency.lockutils [req-22b8b5d4-2f1d-47f8-96f0-c65939176db0 req-9b219914-f1e7-4217-b7a3-2715c44c1381 service nova] Acquired lock "refresh_cache-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.835064] env[61906]: DEBUG nova.network.neutron [req-22b8b5d4-2f1d-47f8-96f0-c65939176db0 req-9b219914-f1e7-4217-b7a3-2715c44c1381 service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Refreshing network info cache for port 3e04c5c7-8d7a-46ae-a2d7-52dc429650f9 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 875.836265] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f1:15:76', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35ac9709-fd8b-4630-897a-68ed629d1b11', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3e04c5c7-8d7a-46ae-a2d7-52dc429650f9', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 875.844334] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Creating folder: Project (e36ff7a500ef444cbf5a168c5a48208b). Parent ref: group-v284713. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 875.847770] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ff2199f9-bce3-4f8f-b0ca-a86890deaba6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.859255] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Created folder: Project (e36ff7a500ef444cbf5a168c5a48208b) in parent group-v284713. [ 875.859528] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Creating folder: Instances. Parent ref: group-v284750. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 875.859811] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aea7450a-9627-4262-b9b8-a71c8abdaacb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.869124] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Created folder: Instances in parent group-v284750. [ 875.869124] env[61906]: DEBUG oslo.service.loopingcall [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.869353] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 875.869586] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-817733eb-ae6a-4ef8-b787-285ebc8947eb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.890978] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 875.890978] env[61906]: value = "task-1333199" [ 875.890978] env[61906]: _type = "Task" [ 875.890978] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.898571] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333199, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.993657] env[61906]: DEBUG nova.compute.manager [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 876.061628] env[61906]: DEBUG nova.compute.manager [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 876.088268] env[61906]: DEBUG nova.virt.hardware [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 876.088522] env[61906]: DEBUG nova.virt.hardware [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 876.088687] env[61906]: DEBUG nova.virt.hardware [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 876.088862] env[61906]: DEBUG nova.virt.hardware [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 876.089011] env[61906]: DEBUG nova.virt.hardware [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 876.089166] env[61906]: DEBUG nova.virt.hardware [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 876.089372] env[61906]: DEBUG nova.virt.hardware [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 876.089530] env[61906]: DEBUG nova.virt.hardware [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 876.089696] env[61906]: DEBUG nova.virt.hardware [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 876.089850] env[61906]: DEBUG nova.virt.hardware [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 876.090028] env[61906]: DEBUG nova.virt.hardware [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 876.091219] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cb8e7f0-b6bb-4557-a860-900c70a1afcc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.094602] env[61906]: DEBUG nova.network.neutron [req-22b8b5d4-2f1d-47f8-96f0-c65939176db0 req-9b219914-f1e7-4217-b7a3-2715c44c1381 service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Updated VIF entry in instance network info cache for port 3e04c5c7-8d7a-46ae-a2d7-52dc429650f9. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 876.094926] env[61906]: DEBUG nova.network.neutron [req-22b8b5d4-2f1d-47f8-96f0-c65939176db0 req-9b219914-f1e7-4217-b7a3-2715c44c1381 service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Updating instance_info_cache with network_info: [{"id": "3e04c5c7-8d7a-46ae-a2d7-52dc429650f9", "address": "fa:16:3e:f1:15:76", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e04c5c7-8d", "ovs_interfaceid": "3e04c5c7-8d7a-46ae-a2d7-52dc429650f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.103040] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3ce2c6c-3f0b-4dee-bb53-a0126a750c5f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.117590] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Instance VIF info [] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 876.123153] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Creating folder: Project (35a0b9ddc77b42eda69d4a74b0d6e70f). Parent ref: group-v284713. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 876.123717] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0bf66a52-6e22-4269-b689-460e59a057ef {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.133655] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Created folder: Project (35a0b9ddc77b42eda69d4a74b0d6e70f) in parent group-v284713. [ 876.133874] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Creating folder: Instances. Parent ref: group-v284753. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 876.134121] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c842e4b5-9ccd-4088-a7a9-3cc522bb387e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.142817] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Created folder: Instances in parent group-v284753. [ 876.143072] env[61906]: DEBUG oslo.service.loopingcall [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 876.143290] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 876.143518] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5c673c96-6555-497e-a7c7-a9321c5ebb62 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.160237] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 876.160237] env[61906]: value = "task-1333202" [ 876.160237] env[61906]: _type = "Task" [ 876.160237] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.167599] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333202, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.257904] env[61906]: DEBUG oslo_vmware.api [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Task: {'id': task-1333195, 'name': ReconfigVM_Task, 'duration_secs': 0.343131} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.258191] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 80a6a0b8-17c3-46e8-8861-6d1759a19111/80a6a0b8-17c3-46e8-8861-6d1759a19111.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.258811] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-38bcf0a1-435b-494e-8ff9-5a155f8a543a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.266577] env[61906]: DEBUG oslo_vmware.api [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Waiting for the task: (returnval){ [ 876.266577] env[61906]: value = "task-1333203" [ 876.266577] env[61906]: _type = "Task" [ 876.266577] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.272378] env[61906]: DEBUG oslo_vmware.api [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333196, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06638} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.272951] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 876.273745] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-998dcd99-4153-41fd-a51a-74ab1b41ffe5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.279128] env[61906]: DEBUG oslo_vmware.api [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Task: {'id': task-1333203, 'name': Rename_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.300712] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] d60575a6-af13-4d81-95d1-081d0f3e2ce6/d60575a6-af13-4d81-95d1-081d0f3e2ce6.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 876.301021] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-349f9c6e-32a6-4681-a9fe-8f5976db7436 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.320912] env[61906]: DEBUG oslo_vmware.api [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Waiting for the task: (returnval){ [ 876.320912] env[61906]: value = "task-1333204" [ 876.320912] env[61906]: _type = "Task" [ 876.320912] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.324405] env[61906]: DEBUG oslo_concurrency.lockutils [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.283s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.324867] env[61906]: DEBUG nova.compute.manager [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 876.327284] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 13.435s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.327459] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.327613] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61906) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 876.327864] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.873s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.329241] env[61906]: INFO nova.compute.claims [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 876.335058] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fbaeda0-ffa9-454b-a5ea-e6e2fa84dcce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.337974] env[61906]: DEBUG oslo_vmware.api [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333204, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.343030] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59dbba2c-93f8-43bd-a46a-d1f858f9c6c1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.358769] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-951f71d1-3c19-4717-a7ff-c0bbfebc91b3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.366679] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d041d6-6f46-40df-984d-0cc5ef989ca0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.404126] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181428MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=61906) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 876.404306] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.413738] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333199, 'name': CreateVM_Task, 'duration_secs': 0.413888} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.413888] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 876.414646] env[61906]: DEBUG oslo_concurrency.lockutils [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.414852] env[61906]: DEBUG oslo_concurrency.lockutils [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.415229] env[61906]: DEBUG oslo_concurrency.lockutils [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 876.415517] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35240dd7-5b49-4a0b-8941-c78a5ba5f27f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.421460] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 876.421460] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]528ade10-f47d-4eb9-89f4-1023bdebe94a" [ 876.421460] env[61906]: _type = "Task" [ 876.421460] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.429320] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]528ade10-f47d-4eb9-89f4-1023bdebe94a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.461460] env[61906]: INFO nova.compute.manager [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Rebuilding instance [ 876.511166] env[61906]: DEBUG nova.compute.manager [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 876.511784] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d9b7b86-98a2-4b32-a20a-8395f4df7089 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.522370] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.598036] env[61906]: DEBUG oslo_concurrency.lockutils [req-22b8b5d4-2f1d-47f8-96f0-c65939176db0 req-9b219914-f1e7-4217-b7a3-2715c44c1381 service nova] Releasing lock "refresh_cache-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.670247] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333202, 'name': CreateVM_Task, 'duration_secs': 0.34033} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.670418] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 876.670851] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.670987] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.671636] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 876.671636] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80feb9aa-e438-4fae-9426-c241097e32c5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.675900] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Waiting for the task: (returnval){ [ 876.675900] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52655204-9d8d-ed80-ad93-09bc5addf921" [ 876.675900] env[61906]: _type = "Task" [ 876.675900] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.683444] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52655204-9d8d-ed80-ad93-09bc5addf921, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.776329] env[61906]: DEBUG oslo_vmware.api [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Task: {'id': task-1333203, 'name': Rename_Task, 'duration_secs': 0.16645} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.776566] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 876.776807] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3776c5ed-0707-478b-ac0c-cde83099962d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.782860] env[61906]: DEBUG oslo_vmware.api [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Waiting for the task: (returnval){ [ 876.782860] env[61906]: value = "task-1333205" [ 876.782860] env[61906]: _type = "Task" [ 876.782860] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.790251] env[61906]: DEBUG oslo_vmware.api [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Task: {'id': task-1333205, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.833158] env[61906]: DEBUG nova.compute.utils [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 876.834459] env[61906]: DEBUG oslo_vmware.api [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333204, 'name': ReconfigVM_Task, 'duration_secs': 0.299652} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.836577] env[61906]: DEBUG nova.compute.manager [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 876.836716] env[61906]: DEBUG nova.network.neutron [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 876.838440] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Reconfigured VM instance instance-00000040 to attach disk [datastore2] d60575a6-af13-4d81-95d1-081d0f3e2ce6/d60575a6-af13-4d81-95d1-081d0f3e2ce6.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.839167] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-17919f00-7783-417d-b72d-c0af2e05fdc8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.845833] env[61906]: DEBUG oslo_vmware.api [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Waiting for the task: (returnval){ [ 876.845833] env[61906]: value = "task-1333206" [ 876.845833] env[61906]: _type = "Task" [ 876.845833] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.856317] env[61906]: DEBUG oslo_vmware.api [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333206, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.918116] env[61906]: DEBUG nova.policy [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ad482253735c467dae34dd54d2d32cc0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5375ff2ae28b4427a3c4541ee3d2729a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 876.933235] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]528ade10-f47d-4eb9-89f4-1023bdebe94a, 'name': SearchDatastore_Task, 'duration_secs': 0.009872} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.933571] env[61906]: DEBUG oslo_concurrency.lockutils [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.933833] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 876.934095] env[61906]: DEBUG oslo_concurrency.lockutils [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.934275] env[61906]: DEBUG oslo_concurrency.lockutils [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.934473] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 876.934770] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6e874397-9a50-4a1c-aad1-37d399c911e7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.943070] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 876.943262] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 876.944018] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-851edf63-adce-480c-abf1-913bb5e7a5b4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.949335] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 876.949335] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52c5ea71-4384-93ca-c727-89f3b684f644" [ 876.949335] env[61906]: _type = "Task" [ 876.949335] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.957647] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52c5ea71-4384-93ca-c727-89f3b684f644, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.024550] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 877.024876] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-31412973-518d-4033-8915-3296b5b8b938 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.034389] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 877.034389] env[61906]: value = "task-1333207" [ 877.034389] env[61906]: _type = "Task" [ 877.034389] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.042696] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333207, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.187281] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52655204-9d8d-ed80-ad93-09bc5addf921, 'name': SearchDatastore_Task, 'duration_secs': 0.010848} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.187667] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.187947] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 877.188272] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.188393] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.188574] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 877.188855] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f38df6a8-6820-44c3-962e-6bdc1b1845cf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.197846] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 877.198096] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 877.198802] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bec784fb-62e9-4af9-a8c8-0ac2eff39964 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.204213] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Waiting for the task: (returnval){ [ 877.204213] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52aae7d4-0867-4e74-7ff7-0a7f1fbf2afb" [ 877.204213] env[61906]: _type = "Task" [ 877.204213] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.212946] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52aae7d4-0867-4e74-7ff7-0a7f1fbf2afb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.261035] env[61906]: DEBUG nova.network.neutron [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Successfully created port: 8bcec3ac-5104-4e72-af50-08fae9dc156c {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 877.292479] env[61906]: DEBUG oslo_vmware.api [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Task: {'id': task-1333205, 'name': PowerOnVM_Task, 'duration_secs': 0.462671} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.292749] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 877.293128] env[61906]: INFO nova.compute.manager [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Took 8.28 seconds to spawn the instance on the hypervisor. [ 877.293282] env[61906]: DEBUG nova.compute.manager [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 877.294154] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2497344-eceb-47e9-9e3d-ac16696d571d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.337641] env[61906]: DEBUG nova.compute.manager [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 877.356529] env[61906]: DEBUG oslo_vmware.api [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333206, 'name': Rename_Task, 'duration_secs': 0.133453} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.356793] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 877.357048] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-81644bd2-fcd3-48c8-827a-6801dd9fe50c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.366022] env[61906]: DEBUG oslo_vmware.api [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Waiting for the task: (returnval){ [ 877.366022] env[61906]: value = "task-1333208" [ 877.366022] env[61906]: _type = "Task" [ 877.366022] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.373661] env[61906]: DEBUG oslo_vmware.api [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333208, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.464664] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52c5ea71-4384-93ca-c727-89f3b684f644, 'name': SearchDatastore_Task, 'duration_secs': 0.009127} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.467592] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99db9696-8df1-4c96-b51a-16316fd948a8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.473203] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 877.473203] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52d7375d-d1af-5950-38f2-01a1676ac6ea" [ 877.473203] env[61906]: _type = "Task" [ 877.473203] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.481997] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52d7375d-d1af-5950-38f2-01a1676ac6ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.545698] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333207, 'name': PowerOffVM_Task, 'duration_secs': 0.200772} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.548237] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 877.548431] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 877.549315] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46848a66-ffd7-4286-9580-5a91df9fa3e4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.555883] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 877.556390] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f513b941-8e47-419d-b222-e7651c2db0b1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.573442] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddc86fd9-b94f-4fda-96f5-04701170e3e1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.580318] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c85cdc8a-e4ed-4b11-9263-6e9065a24d83 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.611609] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2fcb1dd-804a-45fe-a94a-a4775e1462f8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.621379] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3642ec49-bfe4-423a-8edb-a5e827db0da0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.626630] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 877.626849] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 877.627053] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Deleting the datastore file [datastore2] 9f1253e6-4727-4dca-bad2-04f0c5424cdb {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 877.627671] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5f2b30a8-6a9e-4621-83b7-e1eae4dd86bd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.638160] env[61906]: DEBUG nova.compute.provider_tree [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 877.644559] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 877.644559] env[61906]: value = "task-1333210" [ 877.644559] env[61906]: _type = "Task" [ 877.644559] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.653943] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333210, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.714954] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52aae7d4-0867-4e74-7ff7-0a7f1fbf2afb, 'name': SearchDatastore_Task, 'duration_secs': 0.009696} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.715988] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-399fc67b-c710-495b-b6fc-5525a07ba8b9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.722481] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Waiting for the task: (returnval){ [ 877.722481] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52c42c64-a547-a20c-b000-fd927757c629" [ 877.722481] env[61906]: _type = "Task" [ 877.722481] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.730424] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52c42c64-a547-a20c-b000-fd927757c629, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.810945] env[61906]: INFO nova.compute.manager [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Took 35.79 seconds to build instance. [ 877.875330] env[61906]: DEBUG oslo_vmware.api [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333208, 'name': PowerOnVM_Task, 'duration_secs': 0.456544} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.875639] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 877.875889] env[61906]: INFO nova.compute.manager [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Took 6.51 seconds to spawn the instance on the hypervisor. [ 877.876101] env[61906]: DEBUG nova.compute.manager [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 877.877521] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7bdfa4f-5bed-4abb-96ec-79f759ebf1df {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.983177] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52d7375d-d1af-5950-38f2-01a1676ac6ea, 'name': SearchDatastore_Task, 'duration_secs': 0.008688} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.983449] env[61906]: DEBUG oslo_concurrency.lockutils [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.983710] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8/a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 877.983967] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9889ff65-967d-4fec-9748-fa59374fac88 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.990396] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 877.990396] env[61906]: value = "task-1333211" [ 877.990396] env[61906]: _type = "Task" [ 877.990396] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.997955] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333211, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.143491] env[61906]: DEBUG nova.scheduler.client.report [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 878.158738] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333210, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.291578} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.159206] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 878.159511] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 878.159803] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 878.203840] env[61906]: DEBUG nova.compute.manager [req-ee033217-ffe1-4317-a16a-757ad3318a46 req-19718439-2cbc-42f3-9eb3-700eaf9d62fe service nova] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Received event network-changed-b6dc013b-172b-44b2-9f3c-a78cb9245964 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 878.204036] env[61906]: DEBUG nova.compute.manager [req-ee033217-ffe1-4317-a16a-757ad3318a46 req-19718439-2cbc-42f3-9eb3-700eaf9d62fe service nova] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Refreshing instance network info cache due to event network-changed-b6dc013b-172b-44b2-9f3c-a78cb9245964. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 878.204251] env[61906]: DEBUG oslo_concurrency.lockutils [req-ee033217-ffe1-4317-a16a-757ad3318a46 req-19718439-2cbc-42f3-9eb3-700eaf9d62fe service nova] Acquiring lock "refresh_cache-80a6a0b8-17c3-46e8-8861-6d1759a19111" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.204388] env[61906]: DEBUG oslo_concurrency.lockutils [req-ee033217-ffe1-4317-a16a-757ad3318a46 req-19718439-2cbc-42f3-9eb3-700eaf9d62fe service nova] Acquired lock "refresh_cache-80a6a0b8-17c3-46e8-8861-6d1759a19111" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.204542] env[61906]: DEBUG nova.network.neutron [req-ee033217-ffe1-4317-a16a-757ad3318a46 req-19718439-2cbc-42f3-9eb3-700eaf9d62fe service nova] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Refreshing network info cache for port b6dc013b-172b-44b2-9f3c-a78cb9245964 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 878.233677] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52c42c64-a547-a20c-b000-fd927757c629, 'name': SearchDatastore_Task, 'duration_secs': 0.017878} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.233932] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.234202] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8/3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 878.234468] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6961fe35-c4c7-439d-ad5f-b1e5791d180e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.241725] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Waiting for the task: (returnval){ [ 878.241725] env[61906]: value = "task-1333212" [ 878.241725] env[61906]: _type = "Task" [ 878.241725] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.250160] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Task: {'id': task-1333212, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.312685] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46c815e2-beb0-42ac-befc-0d7b25bc55f6 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Lock "80a6a0b8-17c3-46e8-8861-6d1759a19111" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.038s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.351716] env[61906]: DEBUG nova.compute.manager [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 878.376838] env[61906]: DEBUG nova.virt.hardware [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 878.377102] env[61906]: DEBUG nova.virt.hardware [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 878.377267] env[61906]: DEBUG nova.virt.hardware [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 878.377453] env[61906]: DEBUG nova.virt.hardware [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 878.377601] env[61906]: DEBUG nova.virt.hardware [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 878.377750] env[61906]: DEBUG nova.virt.hardware [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 878.377958] env[61906]: DEBUG nova.virt.hardware [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 878.378148] env[61906]: DEBUG nova.virt.hardware [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 878.378326] env[61906]: DEBUG nova.virt.hardware [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 878.378501] env[61906]: DEBUG nova.virt.hardware [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 878.378682] env[61906]: DEBUG nova.virt.hardware [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 878.379573] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5377981-e921-4859-a25e-59b423b2448b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.395499] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c959479-26d8-4f83-afd8-e723192b49a1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.400113] env[61906]: INFO nova.compute.manager [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Took 24.72 seconds to build instance. [ 878.502029] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333211, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.652100] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.324s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.653658] env[61906]: DEBUG nova.compute.manager [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 878.655506] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.362s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.657028] env[61906]: INFO nova.compute.claims [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 878.752482] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Task: {'id': task-1333212, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.815414] env[61906]: DEBUG nova.compute.manager [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 878.890054] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3dfa7d93-d075-4e97-8b4a-2630a5b9007f tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Acquiring lock "80a6a0b8-17c3-46e8-8861-6d1759a19111" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.890054] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3dfa7d93-d075-4e97-8b4a-2630a5b9007f tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Lock "80a6a0b8-17c3-46e8-8861-6d1759a19111" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.890054] env[61906]: INFO nova.compute.manager [None req-3dfa7d93-d075-4e97-8b4a-2630a5b9007f tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Rebooting instance [ 878.902647] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b89362-ec23-4afb-ae88-d4412bdcca8a tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Lock "d60575a6-af13-4d81-95d1-081d0f3e2ce6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.943s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.003394] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333211, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.827377} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.003464] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8/a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 879.003678] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 879.003946] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3d0e27f9-1ad9-409e-8990-efbb580009a9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.014602] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 879.014602] env[61906]: value = "task-1333213" [ 879.014602] env[61906]: _type = "Task" [ 879.014602] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.020934] env[61906]: DEBUG nova.compute.manager [req-39974790-b940-488d-b54f-a082b6babfb7 req-5bf6b639-2f06-4bdd-b151-e10d4c1804b0 service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Received event network-vif-plugged-8bcec3ac-5104-4e72-af50-08fae9dc156c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.021221] env[61906]: DEBUG oslo_concurrency.lockutils [req-39974790-b940-488d-b54f-a082b6babfb7 req-5bf6b639-2f06-4bdd-b151-e10d4c1804b0 service nova] Acquiring lock "75b9738e-e9b6-435a-88bb-851982828d36-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.021470] env[61906]: DEBUG oslo_concurrency.lockutils [req-39974790-b940-488d-b54f-a082b6babfb7 req-5bf6b639-2f06-4bdd-b151-e10d4c1804b0 service nova] Lock "75b9738e-e9b6-435a-88bb-851982828d36-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.021686] env[61906]: DEBUG oslo_concurrency.lockutils [req-39974790-b940-488d-b54f-a082b6babfb7 req-5bf6b639-2f06-4bdd-b151-e10d4c1804b0 service nova] Lock "75b9738e-e9b6-435a-88bb-851982828d36-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.021811] env[61906]: DEBUG nova.compute.manager [req-39974790-b940-488d-b54f-a082b6babfb7 req-5bf6b639-2f06-4bdd-b151-e10d4c1804b0 service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] No waiting events found dispatching network-vif-plugged-8bcec3ac-5104-4e72-af50-08fae9dc156c {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 879.022038] env[61906]: WARNING nova.compute.manager [req-39974790-b940-488d-b54f-a082b6babfb7 req-5bf6b639-2f06-4bdd-b151-e10d4c1804b0 service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Received unexpected event network-vif-plugged-8bcec3ac-5104-4e72-af50-08fae9dc156c for instance with vm_state building and task_state spawning. [ 879.032460] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333213, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.114292] env[61906]: DEBUG nova.network.neutron [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Successfully updated port: 8bcec3ac-5104-4e72-af50-08fae9dc156c {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 879.117297] env[61906]: DEBUG nova.network.neutron [req-ee033217-ffe1-4317-a16a-757ad3318a46 req-19718439-2cbc-42f3-9eb3-700eaf9d62fe service nova] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Updated VIF entry in instance network info cache for port b6dc013b-172b-44b2-9f3c-a78cb9245964. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 879.117651] env[61906]: DEBUG nova.network.neutron [req-ee033217-ffe1-4317-a16a-757ad3318a46 req-19718439-2cbc-42f3-9eb3-700eaf9d62fe service nova] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Updating instance_info_cache with network_info: [{"id": "b6dc013b-172b-44b2-9f3c-a78cb9245964", "address": "fa:16:3e:09:3c:46", "network": {"id": "e9e4d889-5e38-4828-ba70-3ac52f8b72fa", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-389623216-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ac8ee7432c4a414d9d6d7ce5f60ced7a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "822050c7-1845-485d-b87e-73778d21c33c", "external-id": "nsx-vlan-transportzone-701", "segmentation_id": 701, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6dc013b-17", "ovs_interfaceid": "b6dc013b-172b-44b2-9f3c-a78cb9245964", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.161166] env[61906]: DEBUG nova.compute.utils [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 879.165137] env[61906]: DEBUG nova.compute.manager [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 879.165336] env[61906]: DEBUG nova.network.neutron [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 879.196654] env[61906]: DEBUG nova.virt.hardware [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 879.196968] env[61906]: DEBUG nova.virt.hardware [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 879.197086] env[61906]: DEBUG nova.virt.hardware [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 879.197248] env[61906]: DEBUG nova.virt.hardware [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 879.197397] env[61906]: DEBUG nova.virt.hardware [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 879.197584] env[61906]: DEBUG nova.virt.hardware [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 879.197746] env[61906]: DEBUG nova.virt.hardware [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 879.197964] env[61906]: DEBUG nova.virt.hardware [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 879.198083] env[61906]: DEBUG nova.virt.hardware [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 879.198252] env[61906]: DEBUG nova.virt.hardware [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 879.198425] env[61906]: DEBUG nova.virt.hardware [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 879.199313] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6e429fc-99dd-4030-b4dd-5d3d13e10e70 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.204482] env[61906]: DEBUG nova.policy [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b49e5a6aeb1d43e290a645e14861b889', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '53644a4143a24759a9ff2b5e28b84fb4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 879.211191] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f1c5e4f-03ae-4fce-b8e3-f82f9607f1a0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.226483] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ff:cc:9a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '19440099-773e-4a31-b82e-84a4daa5d8fe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '61a42623-e2e5-4245-9560-be57af9a8397', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 879.234782] env[61906]: DEBUG oslo.service.loopingcall [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 879.235909] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 879.235909] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-214a23a4-d27c-4bf8-8fce-e5a65db9f569 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.261137] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Task: {'id': task-1333212, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.808402} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.262438] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8/3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 879.262668] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 879.265026] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 879.265026] env[61906]: value = "task-1333214" [ 879.265026] env[61906]: _type = "Task" [ 879.265026] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.265026] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e5265845-6624-43ab-aae6-b8565dcdc93b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.272629] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333214, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.273832] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Waiting for the task: (returnval){ [ 879.273832] env[61906]: value = "task-1333215" [ 879.273832] env[61906]: _type = "Task" [ 879.273832] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.281725] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Task: {'id': task-1333215, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.343140] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.406046] env[61906]: DEBUG nova.compute.manager [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 879.414918] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3dfa7d93-d075-4e97-8b4a-2630a5b9007f tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Acquiring lock "refresh_cache-80a6a0b8-17c3-46e8-8861-6d1759a19111" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.528822] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333213, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.253009} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.529544] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 879.530718] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce16e299-5a87-4b83-a019-f3f4b7f2c3b7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.557918] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8/a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 879.557918] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9e01753-cf03-43f7-9e7c-0676edc519a4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.576587] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 879.576587] env[61906]: value = "task-1333216" [ 879.576587] env[61906]: _type = "Task" [ 879.576587] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.585505] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333216, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.617461] env[61906]: DEBUG oslo_concurrency.lockutils [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Acquiring lock "refresh_cache-75b9738e-e9b6-435a-88bb-851982828d36" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.617622] env[61906]: DEBUG oslo_concurrency.lockutils [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Acquired lock "refresh_cache-75b9738e-e9b6-435a-88bb-851982828d36" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.617784] env[61906]: DEBUG nova.network.neutron [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 879.619695] env[61906]: DEBUG oslo_concurrency.lockutils [req-ee033217-ffe1-4317-a16a-757ad3318a46 req-19718439-2cbc-42f3-9eb3-700eaf9d62fe service nova] Releasing lock "refresh_cache-80a6a0b8-17c3-46e8-8861-6d1759a19111" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.623021] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3dfa7d93-d075-4e97-8b4a-2630a5b9007f tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Acquired lock "refresh_cache-80a6a0b8-17c3-46e8-8861-6d1759a19111" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.623021] env[61906]: DEBUG nova.network.neutron [None req-3dfa7d93-d075-4e97-8b4a-2630a5b9007f tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 879.633443] env[61906]: DEBUG nova.compute.manager [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 879.634558] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81af03b1-4514-43a7-9372-b17c4c269798 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.669285] env[61906]: DEBUG nova.compute.manager [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 879.675655] env[61906]: DEBUG nova.network.neutron [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Successfully created port: f23a470d-0c5d-4aca-95ea-22c1ad18ca49 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 879.780957] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333214, 'name': CreateVM_Task} progress is 25%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.793356] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Task: {'id': task-1333215, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.25768} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.793624] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 879.794486] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-907669e8-0b61-48e6-a114-6397c60f9c04 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.817657] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8/3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 879.820561] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8602fc60-ce30-4aa8-8dae-4e7225222140 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.842918] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Waiting for the task: (returnval){ [ 879.842918] env[61906]: value = "task-1333217" [ 879.842918] env[61906]: _type = "Task" [ 879.842918] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.854568] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Task: {'id': task-1333217, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.917074] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fea9eed-1f21-4226-86f5-15135caf8ae9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.929038] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ca2236a-ff66-4cf3-85a5-2d4c733e3e2d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.934167] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.965409] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e32d567-db38-4a6f-bd83-ccdb30c63fe4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.973770] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2041031-6a5b-4d78-a978-cf8f65d1c09a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.988836] env[61906]: DEBUG nova.compute.provider_tree [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 880.090950] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333216, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.145088] env[61906]: INFO nova.compute.manager [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] instance snapshotting [ 880.147932] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65757bc0-d7bc-43ee-acb9-50f852cacf48 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.174419] env[61906]: DEBUG nova.network.neutron [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 880.176807] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa10f156-d4d5-47a9-91d5-e93e33762de7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.276302] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333214, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.325531] env[61906]: DEBUG nova.network.neutron [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Updating instance_info_cache with network_info: [{"id": "8bcec3ac-5104-4e72-af50-08fae9dc156c", "address": "fa:16:3e:36:f8:fd", "network": {"id": "eaf64f00-dd3b-41c3-b679-3ee2f53735a0", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-255429219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "5375ff2ae28b4427a3c4541ee3d2729a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20fbc86c-0935-406c-9659-3ffc33fc0d08", "external-id": "nsx-vlan-transportzone-191", "segmentation_id": 191, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bcec3ac-51", "ovs_interfaceid": "8bcec3ac-5104-4e72-af50-08fae9dc156c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.353387] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Task: {'id': task-1333217, 'name': ReconfigVM_Task, 'duration_secs': 0.369097} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.353659] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Reconfigured VM instance instance-00000042 to attach disk [datastore1] 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8/3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.354284] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a0540a5e-0d60-4f51-9575-858ff56c8942 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.361106] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Waiting for the task: (returnval){ [ 880.361106] env[61906]: value = "task-1333218" [ 880.361106] env[61906]: _type = "Task" [ 880.361106] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.364881] env[61906]: DEBUG nova.network.neutron [None req-3dfa7d93-d075-4e97-8b4a-2630a5b9007f tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Updating instance_info_cache with network_info: [{"id": "b6dc013b-172b-44b2-9f3c-a78cb9245964", "address": "fa:16:3e:09:3c:46", "network": {"id": "e9e4d889-5e38-4828-ba70-3ac52f8b72fa", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-389623216-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ac8ee7432c4a414d9d6d7ce5f60ced7a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "822050c7-1845-485d-b87e-73778d21c33c", "external-id": "nsx-vlan-transportzone-701", "segmentation_id": 701, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6dc013b-17", "ovs_interfaceid": "b6dc013b-172b-44b2-9f3c-a78cb9245964", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.370956] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Task: {'id': task-1333218, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.493026] env[61906]: DEBUG nova.scheduler.client.report [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 880.587249] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333216, 'name': ReconfigVM_Task, 'duration_secs': 0.53015} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.587538] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Reconfigured VM instance instance-00000041 to attach disk [datastore2] a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8/a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.588194] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fff5b5de-991c-4b55-a7a3-de45b0082d07 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.594211] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 880.594211] env[61906]: value = "task-1333219" [ 880.594211] env[61906]: _type = "Task" [ 880.594211] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.603142] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333219, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.686018] env[61906]: DEBUG nova.compute.manager [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 880.691223] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Creating Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 880.691525] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b7cebc5d-a86b-4d84-89e8-857fa7fd50e7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.699786] env[61906]: DEBUG oslo_vmware.api [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Waiting for the task: (returnval){ [ 880.699786] env[61906]: value = "task-1333220" [ 880.699786] env[61906]: _type = "Task" [ 880.699786] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.708386] env[61906]: DEBUG nova.virt.hardware [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 880.708625] env[61906]: DEBUG nova.virt.hardware [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 880.708784] env[61906]: DEBUG nova.virt.hardware [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 880.708973] env[61906]: DEBUG nova.virt.hardware [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 880.709129] env[61906]: DEBUG nova.virt.hardware [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 880.709279] env[61906]: DEBUG nova.virt.hardware [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 880.709522] env[61906]: DEBUG nova.virt.hardware [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 880.709716] env[61906]: DEBUG nova.virt.hardware [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 880.709801] env[61906]: DEBUG nova.virt.hardware [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 880.709954] env[61906]: DEBUG nova.virt.hardware [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 880.710144] env[61906]: DEBUG nova.virt.hardware [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 880.710941] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7dfba98-0f22-417a-960a-c673d5cfac2f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.716328] env[61906]: DEBUG oslo_vmware.api [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333220, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.722454] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a939988e-c227-4f53-8921-45f3af2e4bf7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.776191] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333214, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.828428] env[61906]: DEBUG oslo_concurrency.lockutils [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Releasing lock "refresh_cache-75b9738e-e9b6-435a-88bb-851982828d36" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.828776] env[61906]: DEBUG nova.compute.manager [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Instance network_info: |[{"id": "8bcec3ac-5104-4e72-af50-08fae9dc156c", "address": "fa:16:3e:36:f8:fd", "network": {"id": "eaf64f00-dd3b-41c3-b679-3ee2f53735a0", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-255429219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "5375ff2ae28b4427a3c4541ee3d2729a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20fbc86c-0935-406c-9659-3ffc33fc0d08", "external-id": "nsx-vlan-transportzone-191", "segmentation_id": 191, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bcec3ac-51", "ovs_interfaceid": "8bcec3ac-5104-4e72-af50-08fae9dc156c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 880.829231] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:f8:fd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '20fbc86c-0935-406c-9659-3ffc33fc0d08', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8bcec3ac-5104-4e72-af50-08fae9dc156c', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 880.836842] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Creating folder: Project (5375ff2ae28b4427a3c4541ee3d2729a). Parent ref: group-v284713. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 880.837157] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-76cc7ce7-bf92-4f4f-b41c-f44c61f4aace {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.848099] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Created folder: Project (5375ff2ae28b4427a3c4541ee3d2729a) in parent group-v284713. [ 880.848099] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Creating folder: Instances. Parent ref: group-v284757. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 880.848193] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e8e8c23e-e333-4251-981d-8525d29d647e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.856870] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Created folder: Instances in parent group-v284757. [ 880.857113] env[61906]: DEBUG oslo.service.loopingcall [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.857308] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 880.857508] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6e3224b1-9315-49fa-9161-a4d6f30c27dc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.871531] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3dfa7d93-d075-4e97-8b4a-2630a5b9007f tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Releasing lock "refresh_cache-80a6a0b8-17c3-46e8-8861-6d1759a19111" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.876619] env[61906]: DEBUG nova.compute.manager [None req-3dfa7d93-d075-4e97-8b4a-2630a5b9007f tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 880.877689] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63c1d66c-0347-44ea-a5e8-077e2e614fce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.881744] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 880.881744] env[61906]: value = "task-1333223" [ 880.881744] env[61906]: _type = "Task" [ 880.881744] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.889439] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Task: {'id': task-1333218, 'name': Rename_Task, 'duration_secs': 0.186076} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.892199] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 880.892648] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4855e3d6-fd39-4fb4-9383-76334f67aac3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.897301] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333223, 'name': CreateVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.902375] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Waiting for the task: (returnval){ [ 880.902375] env[61906]: value = "task-1333224" [ 880.902375] env[61906]: _type = "Task" [ 880.902375] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.910687] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Task: {'id': task-1333224, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.999040] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.343s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.999666] env[61906]: DEBUG nova.compute.manager [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 881.003455] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.177s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.005292] env[61906]: INFO nova.compute.claims [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 881.105728] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333219, 'name': Rename_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.113008] env[61906]: DEBUG nova.compute.manager [req-1a9a4162-ba4b-4f1c-b41c-7165f7a407b0 req-14d30ce7-164a-45a4-a898-46ffa35853ee service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Received event network-changed-8bcec3ac-5104-4e72-af50-08fae9dc156c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 881.113206] env[61906]: DEBUG nova.compute.manager [req-1a9a4162-ba4b-4f1c-b41c-7165f7a407b0 req-14d30ce7-164a-45a4-a898-46ffa35853ee service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Refreshing instance network info cache due to event network-changed-8bcec3ac-5104-4e72-af50-08fae9dc156c. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 881.113428] env[61906]: DEBUG oslo_concurrency.lockutils [req-1a9a4162-ba4b-4f1c-b41c-7165f7a407b0 req-14d30ce7-164a-45a4-a898-46ffa35853ee service nova] Acquiring lock "refresh_cache-75b9738e-e9b6-435a-88bb-851982828d36" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.113557] env[61906]: DEBUG oslo_concurrency.lockutils [req-1a9a4162-ba4b-4f1c-b41c-7165f7a407b0 req-14d30ce7-164a-45a4-a898-46ffa35853ee service nova] Acquired lock "refresh_cache-75b9738e-e9b6-435a-88bb-851982828d36" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.113719] env[61906]: DEBUG nova.network.neutron [req-1a9a4162-ba4b-4f1c-b41c-7165f7a407b0 req-14d30ce7-164a-45a4-a898-46ffa35853ee service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Refreshing network info cache for port 8bcec3ac-5104-4e72-af50-08fae9dc156c {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 881.211477] env[61906]: DEBUG oslo_vmware.api [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333220, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.276954] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333214, 'name': CreateVM_Task, 'duration_secs': 1.76565} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.280794] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 881.280794] env[61906]: DEBUG oslo_concurrency.lockutils [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.280794] env[61906]: DEBUG oslo_concurrency.lockutils [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.280794] env[61906]: DEBUG oslo_concurrency.lockutils [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 881.280794] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33ad0bae-152c-40cb-9391-0d96df8285c0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.287657] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 881.287657] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52ae90d9-049f-8c68-55ba-f3acbda16f26" [ 881.287657] env[61906]: _type = "Task" [ 881.287657] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.306029] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52ae90d9-049f-8c68-55ba-f3acbda16f26, 'name': SearchDatastore_Task, 'duration_secs': 0.01064} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.306029] env[61906]: DEBUG oslo_concurrency.lockutils [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.306029] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 881.306029] env[61906]: DEBUG oslo_concurrency.lockutils [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.306029] env[61906]: DEBUG oslo_concurrency.lockutils [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.306029] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 881.306029] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-41cc9d60-7a42-451c-a049-a639e3628d17 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.316020] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 881.316020] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 881.316020] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-926e94ed-a756-46ba-8ae8-0caae0e2e1fa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.323284] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 881.323284] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5204b744-4ec7-308f-01ae-53988ac2e876" [ 881.323284] env[61906]: _type = "Task" [ 881.323284] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.333135] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5204b744-4ec7-308f-01ae-53988ac2e876, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.396545] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333223, 'name': CreateVM_Task, 'duration_secs': 0.328942} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.396990] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 881.397729] env[61906]: DEBUG oslo_concurrency.lockutils [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.397893] env[61906]: DEBUG oslo_concurrency.lockutils [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.398239] env[61906]: DEBUG oslo_concurrency.lockutils [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 881.398501] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6ca42ff-007e-4b7e-9b03-e19ed4f6dab9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.408289] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Waiting for the task: (returnval){ [ 881.408289] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5280e130-a8e6-5785-0f5f-c9e077ba7fa8" [ 881.408289] env[61906]: _type = "Task" [ 881.408289] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.411562] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Task: {'id': task-1333224, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.419169] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5280e130-a8e6-5785-0f5f-c9e077ba7fa8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.430404] env[61906]: DEBUG nova.network.neutron [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Successfully updated port: f23a470d-0c5d-4aca-95ea-22c1ad18ca49 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 881.511594] env[61906]: DEBUG nova.compute.utils [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 881.512986] env[61906]: DEBUG nova.compute.manager [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 881.513164] env[61906]: DEBUG nova.network.neutron [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 881.553119] env[61906]: DEBUG nova.policy [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc6b374325bc4ef9882e5f84ddd62fd1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f2d2198383d74dac90aa9727a62176df', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 881.605648] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333219, 'name': Rename_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.711896] env[61906]: DEBUG oslo_vmware.api [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333220, 'name': CreateSnapshot_Task, 'duration_secs': 0.852811} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.714240] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Created Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 881.715279] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d80e2523-2f2d-4dea-8f9a-0e6bdf9f114d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.808064] env[61906]: DEBUG nova.network.neutron [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Successfully created port: 58de857e-fc72-4f45-9da6-93623ceb23bc {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 881.822053] env[61906]: DEBUG nova.network.neutron [req-1a9a4162-ba4b-4f1c-b41c-7165f7a407b0 req-14d30ce7-164a-45a4-a898-46ffa35853ee service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Updated VIF entry in instance network info cache for port 8bcec3ac-5104-4e72-af50-08fae9dc156c. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 881.822458] env[61906]: DEBUG nova.network.neutron [req-1a9a4162-ba4b-4f1c-b41c-7165f7a407b0 req-14d30ce7-164a-45a4-a898-46ffa35853ee service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Updating instance_info_cache with network_info: [{"id": "8bcec3ac-5104-4e72-af50-08fae9dc156c", "address": "fa:16:3e:36:f8:fd", "network": {"id": "eaf64f00-dd3b-41c3-b679-3ee2f53735a0", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-255429219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "5375ff2ae28b4427a3c4541ee3d2729a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20fbc86c-0935-406c-9659-3ffc33fc0d08", "external-id": "nsx-vlan-transportzone-191", "segmentation_id": 191, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bcec3ac-51", "ovs_interfaceid": "8bcec3ac-5104-4e72-af50-08fae9dc156c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.838750] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5204b744-4ec7-308f-01ae-53988ac2e876, 'name': SearchDatastore_Task, 'duration_secs': 0.009737} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.840076] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec10df71-2b67-4381-bad0-deea4d6420b7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.846056] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 881.846056] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52a0adc5-fd00-bf19-ab1d-4a96d58741aa" [ 881.846056] env[61906]: _type = "Task" [ 881.846056] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.854352] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52a0adc5-fd00-bf19-ab1d-4a96d58741aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.899077] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeda35ad-158f-459d-93e3-3a230d570fc2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.909965] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3dfa7d93-d075-4e97-8b4a-2630a5b9007f tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Doing hard reboot of VM {{(pid=61906) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 881.910621] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-e0c85196-b3a5-4ffe-b7b3-66dcf7011d81 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.918962] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Task: {'id': task-1333224, 'name': PowerOnVM_Task} progress is 91%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.920716] env[61906]: DEBUG oslo_vmware.api [None req-3dfa7d93-d075-4e97-8b4a-2630a5b9007f tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Waiting for the task: (returnval){ [ 881.920716] env[61906]: value = "task-1333225" [ 881.920716] env[61906]: _type = "Task" [ 881.920716] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.928538] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5280e130-a8e6-5785-0f5f-c9e077ba7fa8, 'name': SearchDatastore_Task, 'duration_secs': 0.028308} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.929169] env[61906]: DEBUG oslo_concurrency.lockutils [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.929440] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 881.929636] env[61906]: DEBUG oslo_concurrency.lockutils [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.933412] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "refresh_cache-cb34a0ce-d274-4167-8367-e19809cc9e4d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.933559] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired lock "refresh_cache-cb34a0ce-d274-4167-8367-e19809cc9e4d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.933668] env[61906]: DEBUG nova.network.neutron [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 881.934756] env[61906]: DEBUG oslo_vmware.api [None req-3dfa7d93-d075-4e97-8b4a-2630a5b9007f tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Task: {'id': task-1333225, 'name': ResetVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.017454] env[61906]: DEBUG nova.compute.manager [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 882.107723] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333219, 'name': Rename_Task, 'duration_secs': 1.209824} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.107919] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 882.108175] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6dc6064a-7c17-4422-ab49-10364f3bd208 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.117085] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 882.117085] env[61906]: value = "task-1333226" [ 882.117085] env[61906]: _type = "Task" [ 882.117085] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.127215] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333226, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.195556] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-999bf2b0-99ee-4d89-9e36-b8ba804de8d9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.202937] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c972c0f2-5619-48b0-82c3-c3b5ccc5d0a9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.239422] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Creating linked-clone VM from snapshot {{(pid=61906) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 882.239979] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-7db552d4-381f-4636-9af6-76682b0157ff {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.243438] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f98f8a09-7270-4290-9719-41c7699c8655 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.252846] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64cd818d-b969-4dd0-bca8-46f2c7afd821 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.256465] env[61906]: DEBUG oslo_vmware.api [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Waiting for the task: (returnval){ [ 882.256465] env[61906]: value = "task-1333227" [ 882.256465] env[61906]: _type = "Task" [ 882.256465] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.267054] env[61906]: DEBUG nova.compute.provider_tree [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 882.276218] env[61906]: DEBUG oslo_vmware.api [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333227, 'name': CloneVM_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.334024] env[61906]: DEBUG oslo_concurrency.lockutils [req-1a9a4162-ba4b-4f1c-b41c-7165f7a407b0 req-14d30ce7-164a-45a4-a898-46ffa35853ee service nova] Releasing lock "refresh_cache-75b9738e-e9b6-435a-88bb-851982828d36" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.357950] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52a0adc5-fd00-bf19-ab1d-4a96d58741aa, 'name': SearchDatastore_Task, 'duration_secs': 0.023365} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.358244] env[61906]: DEBUG oslo_concurrency.lockutils [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.358570] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 9f1253e6-4727-4dca-bad2-04f0c5424cdb/9f1253e6-4727-4dca-bad2-04f0c5424cdb.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 882.358955] env[61906]: DEBUG oslo_concurrency.lockutils [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.359180] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 882.360034] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d155b226-426e-497c-8a25-f203476b7611 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.361456] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5beca974-8375-49b6-b89e-02e2c098bc4d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.367985] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 882.367985] env[61906]: value = "task-1333228" [ 882.367985] env[61906]: _type = "Task" [ 882.367985] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.372282] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 882.372476] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 882.373568] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad0b809a-cf1b-4b29-836f-c72d9285f960 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.379025] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333228, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.382148] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Waiting for the task: (returnval){ [ 882.382148] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52297c50-cf82-f9d8-98f8-60d786eba6fc" [ 882.382148] env[61906]: _type = "Task" [ 882.382148] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.390836] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52297c50-cf82-f9d8-98f8-60d786eba6fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.413321] env[61906]: DEBUG oslo_vmware.api [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Task: {'id': task-1333224, 'name': PowerOnVM_Task, 'duration_secs': 1.022278} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.413622] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 882.413830] env[61906]: INFO nova.compute.manager [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Took 6.35 seconds to spawn the instance on the hypervisor. [ 882.414015] env[61906]: DEBUG nova.compute.manager [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 882.415032] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d127124a-9c67-4142-bec0-8542729de08f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.431574] env[61906]: DEBUG oslo_vmware.api [None req-3dfa7d93-d075-4e97-8b4a-2630a5b9007f tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Task: {'id': task-1333225, 'name': ResetVM_Task, 'duration_secs': 0.09142} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.431925] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3dfa7d93-d075-4e97-8b4a-2630a5b9007f tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Did hard reboot of VM {{(pid=61906) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 882.432143] env[61906]: DEBUG nova.compute.manager [None req-3dfa7d93-d075-4e97-8b4a-2630a5b9007f tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 882.432936] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36f7dbc9-913a-4521-b522-dce80fac8e14 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.470180] env[61906]: DEBUG nova.network.neutron [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 882.628419] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333226, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.644460] env[61906]: DEBUG nova.network.neutron [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Updating instance_info_cache with network_info: [{"id": "f23a470d-0c5d-4aca-95ea-22c1ad18ca49", "address": "fa:16:3e:59:77:5b", "network": {"id": "bc9185e9-f26b-489c-84ee-e6db41ac6283", "bridge": "br-int", "label": "tempest-ServersTestJSON-1167747148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53644a4143a24759a9ff2b5e28b84fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf23a470d-0c", "ovs_interfaceid": "f23a470d-0c5d-4aca-95ea-22c1ad18ca49", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.768694] env[61906]: DEBUG oslo_vmware.api [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333227, 'name': CloneVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.769732] env[61906]: DEBUG nova.scheduler.client.report [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 882.880251] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333228, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.892816] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52297c50-cf82-f9d8-98f8-60d786eba6fc, 'name': SearchDatastore_Task, 'duration_secs': 0.010633} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.893856] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2b20d6e-9bc8-4cdc-b180-b6d05079be97 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.901221] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Waiting for the task: (returnval){ [ 882.901221] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52b5a850-6c18-dad7-3dae-36661f0b5981" [ 882.901221] env[61906]: _type = "Task" [ 882.901221] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.911744] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52b5a850-6c18-dad7-3dae-36661f0b5981, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.934856] env[61906]: INFO nova.compute.manager [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Took 23.28 seconds to build instance. [ 882.948590] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3dfa7d93-d075-4e97-8b4a-2630a5b9007f tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Lock "80a6a0b8-17c3-46e8-8861-6d1759a19111" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.061s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.028207] env[61906]: DEBUG nova.compute.manager [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 883.050199] env[61906]: DEBUG nova.virt.hardware [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 883.050452] env[61906]: DEBUG nova.virt.hardware [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 883.050612] env[61906]: DEBUG nova.virt.hardware [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 883.050787] env[61906]: DEBUG nova.virt.hardware [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 883.050933] env[61906]: DEBUG nova.virt.hardware [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 883.051091] env[61906]: DEBUG nova.virt.hardware [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 883.051303] env[61906]: DEBUG nova.virt.hardware [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 883.051463] env[61906]: DEBUG nova.virt.hardware [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 883.051627] env[61906]: DEBUG nova.virt.hardware [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 883.051788] env[61906]: DEBUG nova.virt.hardware [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 883.051958] env[61906]: DEBUG nova.virt.hardware [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 883.052846] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c517173b-9289-4d9a-8c55-91584875198b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.060111] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb703f0b-6dfe-4423-90e5-23bdada6e07a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.127189] env[61906]: DEBUG oslo_vmware.api [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333226, 'name': PowerOnVM_Task, 'duration_secs': 0.655126} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.127546] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 883.127662] env[61906]: INFO nova.compute.manager [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Took 9.38 seconds to spawn the instance on the hypervisor. [ 883.127832] env[61906]: DEBUG nova.compute.manager [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 883.128609] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-545ce7c4-c637-412b-82c7-115eca39e81e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.141023] env[61906]: DEBUG nova.compute.manager [req-17ef7136-8d3a-44c6-92f8-2df96d558fa6 req-0a7b218d-b77b-4e06-98c5-f4ef6c5b0b0f service nova] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Received event network-vif-plugged-f23a470d-0c5d-4aca-95ea-22c1ad18ca49 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 883.141023] env[61906]: DEBUG oslo_concurrency.lockutils [req-17ef7136-8d3a-44c6-92f8-2df96d558fa6 req-0a7b218d-b77b-4e06-98c5-f4ef6c5b0b0f service nova] Acquiring lock "cb34a0ce-d274-4167-8367-e19809cc9e4d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.141023] env[61906]: DEBUG oslo_concurrency.lockutils [req-17ef7136-8d3a-44c6-92f8-2df96d558fa6 req-0a7b218d-b77b-4e06-98c5-f4ef6c5b0b0f service nova] Lock "cb34a0ce-d274-4167-8367-e19809cc9e4d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.141023] env[61906]: DEBUG oslo_concurrency.lockutils [req-17ef7136-8d3a-44c6-92f8-2df96d558fa6 req-0a7b218d-b77b-4e06-98c5-f4ef6c5b0b0f service nova] Lock "cb34a0ce-d274-4167-8367-e19809cc9e4d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.141755] env[61906]: DEBUG nova.compute.manager [req-17ef7136-8d3a-44c6-92f8-2df96d558fa6 req-0a7b218d-b77b-4e06-98c5-f4ef6c5b0b0f service nova] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] No waiting events found dispatching network-vif-plugged-f23a470d-0c5d-4aca-95ea-22c1ad18ca49 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 883.141755] env[61906]: WARNING nova.compute.manager [req-17ef7136-8d3a-44c6-92f8-2df96d558fa6 req-0a7b218d-b77b-4e06-98c5-f4ef6c5b0b0f service nova] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Received unexpected event network-vif-plugged-f23a470d-0c5d-4aca-95ea-22c1ad18ca49 for instance with vm_state building and task_state spawning. [ 883.141755] env[61906]: DEBUG nova.compute.manager [req-17ef7136-8d3a-44c6-92f8-2df96d558fa6 req-0a7b218d-b77b-4e06-98c5-f4ef6c5b0b0f service nova] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Received event network-changed-f23a470d-0c5d-4aca-95ea-22c1ad18ca49 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 883.141755] env[61906]: DEBUG nova.compute.manager [req-17ef7136-8d3a-44c6-92f8-2df96d558fa6 req-0a7b218d-b77b-4e06-98c5-f4ef6c5b0b0f service nova] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Refreshing instance network info cache due to event network-changed-f23a470d-0c5d-4aca-95ea-22c1ad18ca49. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 883.141949] env[61906]: DEBUG oslo_concurrency.lockutils [req-17ef7136-8d3a-44c6-92f8-2df96d558fa6 req-0a7b218d-b77b-4e06-98c5-f4ef6c5b0b0f service nova] Acquiring lock "refresh_cache-cb34a0ce-d274-4167-8367-e19809cc9e4d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.146817] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Releasing lock "refresh_cache-cb34a0ce-d274-4167-8367-e19809cc9e4d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.147168] env[61906]: DEBUG nova.compute.manager [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Instance network_info: |[{"id": "f23a470d-0c5d-4aca-95ea-22c1ad18ca49", "address": "fa:16:3e:59:77:5b", "network": {"id": "bc9185e9-f26b-489c-84ee-e6db41ac6283", "bridge": "br-int", "label": "tempest-ServersTestJSON-1167747148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53644a4143a24759a9ff2b5e28b84fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf23a470d-0c", "ovs_interfaceid": "f23a470d-0c5d-4aca-95ea-22c1ad18ca49", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 883.147451] env[61906]: DEBUG oslo_concurrency.lockutils [req-17ef7136-8d3a-44c6-92f8-2df96d558fa6 req-0a7b218d-b77b-4e06-98c5-f4ef6c5b0b0f service nova] Acquired lock "refresh_cache-cb34a0ce-d274-4167-8367-e19809cc9e4d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.147647] env[61906]: DEBUG nova.network.neutron [req-17ef7136-8d3a-44c6-92f8-2df96d558fa6 req-0a7b218d-b77b-4e06-98c5-f4ef6c5b0b0f service nova] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Refreshing network info cache for port f23a470d-0c5d-4aca-95ea-22c1ad18ca49 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 883.149026] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:59:77:5b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f23a470d-0c5d-4aca-95ea-22c1ad18ca49', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 883.156257] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Creating folder: Project (53644a4143a24759a9ff2b5e28b84fb4). Parent ref: group-v284713. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 883.159600] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0ad1628f-07af-49cc-91c1-5af379051bed {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.170512] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Created folder: Project (53644a4143a24759a9ff2b5e28b84fb4) in parent group-v284713. [ 883.170512] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Creating folder: Instances. Parent ref: group-v284762. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 883.171111] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e157baef-60fd-4769-accb-4320d6f34a21 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.179760] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Created folder: Instances in parent group-v284762. [ 883.180019] env[61906]: DEBUG oslo.service.loopingcall [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.180212] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 883.180415] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-74d52ef9-15ee-4b11-901b-e8746fb06003 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.203718] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 883.203718] env[61906]: value = "task-1333231" [ 883.203718] env[61906]: _type = "Task" [ 883.203718] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.215830] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333231, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.269051] env[61906]: DEBUG oslo_vmware.api [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333227, 'name': CloneVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.274984] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.272s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.275573] env[61906]: DEBUG nova.compute.manager [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 883.278603] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.164s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.280505] env[61906]: INFO nova.compute.claims [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 883.379609] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333228, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.60168} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.379891] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 9f1253e6-4727-4dca-bad2-04f0c5424cdb/9f1253e6-4727-4dca-bad2-04f0c5424cdb.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 883.380179] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 883.380443] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7fb5371d-f95a-4b33-b5e6-bb435878bc11 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.386727] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 883.386727] env[61906]: value = "task-1333232" [ 883.386727] env[61906]: _type = "Task" [ 883.386727] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.396057] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333232, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.411363] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52b5a850-6c18-dad7-3dae-36661f0b5981, 'name': SearchDatastore_Task, 'duration_secs': 0.018904} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.411626] env[61906]: DEBUG oslo_concurrency.lockutils [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.411880] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 75b9738e-e9b6-435a-88bb-851982828d36/75b9738e-e9b6-435a-88bb-851982828d36.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 883.412185] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fd261b7a-b5e0-4712-aba1-0d5c37321c2d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.419027] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Waiting for the task: (returnval){ [ 883.419027] env[61906]: value = "task-1333233" [ 883.419027] env[61906]: _type = "Task" [ 883.419027] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.429394] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333233, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.437138] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a1e3615-8aa1-4630-8742-1481727a4e85 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Lock "3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.404s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.444487] env[61906]: DEBUG nova.network.neutron [req-17ef7136-8d3a-44c6-92f8-2df96d558fa6 req-0a7b218d-b77b-4e06-98c5-f4ef6c5b0b0f service nova] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Updated VIF entry in instance network info cache for port f23a470d-0c5d-4aca-95ea-22c1ad18ca49. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 883.444849] env[61906]: DEBUG nova.network.neutron [req-17ef7136-8d3a-44c6-92f8-2df96d558fa6 req-0a7b218d-b77b-4e06-98c5-f4ef6c5b0b0f service nova] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Updating instance_info_cache with network_info: [{"id": "f23a470d-0c5d-4aca-95ea-22c1ad18ca49", "address": "fa:16:3e:59:77:5b", "network": {"id": "bc9185e9-f26b-489c-84ee-e6db41ac6283", "bridge": "br-int", "label": "tempest-ServersTestJSON-1167747148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53644a4143a24759a9ff2b5e28b84fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf23a470d-0c", "ovs_interfaceid": "f23a470d-0c5d-4aca-95ea-22c1ad18ca49", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.540700] env[61906]: DEBUG nova.network.neutron [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Successfully updated port: 58de857e-fc72-4f45-9da6-93623ceb23bc {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 883.647094] env[61906]: INFO nova.compute.manager [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Took 25.72 seconds to build instance. [ 883.713649] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333231, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.769269] env[61906]: DEBUG oslo_vmware.api [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333227, 'name': CloneVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.785847] env[61906]: DEBUG nova.compute.utils [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 883.789303] env[61906]: DEBUG nova.compute.manager [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 883.789525] env[61906]: DEBUG nova.network.neutron [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 883.890177] env[61906]: DEBUG nova.policy [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1b017cdae2af4ae8aa069ae23e744797', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '129cdb06b47346b6895df45e73f7b2dd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 883.901334] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333232, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066193} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.901943] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 883.903290] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a00fa15f-867e-4328-aeff-18246c3a0c27 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.936538] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] 9f1253e6-4727-4dca-bad2-04f0c5424cdb/9f1253e6-4727-4dca-bad2-04f0c5424cdb.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 883.940291] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a1677ed2-a915-4733-afbe-7f167cba9a95 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.958848] env[61906]: DEBUG oslo_concurrency.lockutils [req-17ef7136-8d3a-44c6-92f8-2df96d558fa6 req-0a7b218d-b77b-4e06-98c5-f4ef6c5b0b0f service nova] Releasing lock "refresh_cache-cb34a0ce-d274-4167-8367-e19809cc9e4d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.965111] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333233, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.967181] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 883.967181] env[61906]: value = "task-1333234" [ 883.967181] env[61906]: _type = "Task" [ 883.967181] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.977898] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333234, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.043243] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "refresh_cache-cc8870d0-eafb-4f73-aa79-b98f51370237" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.043552] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquired lock "refresh_cache-cc8870d0-eafb-4f73-aa79-b98f51370237" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.043736] env[61906]: DEBUG nova.network.neutron [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 884.149621] env[61906]: DEBUG oslo_concurrency.lockutils [None req-35ced531-d234-4eba-af14-a6daaae99aad tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.762s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.213218] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333231, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.269653] env[61906]: DEBUG oslo_vmware.api [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333227, 'name': CloneVM_Task, 'duration_secs': 1.916087} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.269907] env[61906]: INFO nova.virt.vmwareapi.vmops [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Created linked-clone VM from snapshot [ 884.270696] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a42c7035-6762-44b0-9fa3-e956f0c11efc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.281169] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Uploading image 9ea4f3cc-c567-4887-b2ca-8874462a07fc {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 884.290180] env[61906]: DEBUG nova.compute.manager [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 884.309923] env[61906]: DEBUG oslo_vmware.rw_handles [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 884.309923] env[61906]: value = "vm-284761" [ 884.309923] env[61906]: _type = "VirtualMachine" [ 884.309923] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 884.311276] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-951508ac-1af2-4990-bd47-edc788b717dc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.320657] env[61906]: DEBUG oslo_vmware.rw_handles [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Lease: (returnval){ [ 884.320657] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52b498e9-7a4b-b171-2530-ab211162ff9b" [ 884.320657] env[61906]: _type = "HttpNfcLease" [ 884.320657] env[61906]: } obtained for exporting VM: (result){ [ 884.320657] env[61906]: value = "vm-284761" [ 884.320657] env[61906]: _type = "VirtualMachine" [ 884.320657] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 884.320906] env[61906]: DEBUG oslo_vmware.api [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Waiting for the lease: (returnval){ [ 884.320906] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52b498e9-7a4b-b171-2530-ab211162ff9b" [ 884.320906] env[61906]: _type = "HttpNfcLease" [ 884.320906] env[61906]: } to be ready. {{(pid=61906) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 884.330809] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 884.330809] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52b498e9-7a4b-b171-2530-ab211162ff9b" [ 884.330809] env[61906]: _type = "HttpNfcLease" [ 884.330809] env[61906]: } is initializing. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 884.442021] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333233, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.789205} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.442021] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 75b9738e-e9b6-435a-88bb-851982828d36/75b9738e-e9b6-435a-88bb-851982828d36.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 884.442021] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 884.442021] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8523903d-2199-4fb5-94f5-4a75077b73e1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.450662] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Waiting for the task: (returnval){ [ 884.450662] env[61906]: value = "task-1333236" [ 884.450662] env[61906]: _type = "Task" [ 884.450662] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.460607] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333236, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.477655] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333234, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.524110] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61dd3609-9137-404d-8a8a-6c2de1a9ff8e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.532465] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7094c460-562a-442d-bda4-b31fafabaa2b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.566415] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71caa3f7-fe46-4fe9-a6b9-0350d9ae3c18 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.576017] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Acquiring lock "80a6a0b8-17c3-46e8-8861-6d1759a19111" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.576017] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Lock "80a6a0b8-17c3-46e8-8861-6d1759a19111" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.576017] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Acquiring lock "80a6a0b8-17c3-46e8-8861-6d1759a19111-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.576017] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Lock "80a6a0b8-17c3-46e8-8861-6d1759a19111-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.576017] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Lock "80a6a0b8-17c3-46e8-8861-6d1759a19111-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.578300] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42ce9df9-2c0a-4112-996a-d4bb1a4c2915 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.587258] env[61906]: INFO nova.compute.manager [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Terminating instance [ 884.587258] env[61906]: DEBUG nova.compute.manager [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 884.587258] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 884.587571] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f17949e1-28c6-4587-8a75-6504d1f3aaae {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.597498] env[61906]: DEBUG nova.compute.provider_tree [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.604116] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 884.604116] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e85f112c-f4c5-4a8d-84f7-475750fd6678 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.606347] env[61906]: DEBUG nova.network.neutron [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 884.610116] env[61906]: DEBUG oslo_vmware.api [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Waiting for the task: (returnval){ [ 884.610116] env[61906]: value = "task-1333237" [ 884.610116] env[61906]: _type = "Task" [ 884.610116] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.618278] env[61906]: DEBUG oslo_vmware.api [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Task: {'id': task-1333237, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.714981] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333231, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.716799] env[61906]: DEBUG nova.network.neutron [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Successfully created port: 175a521b-1c70-41f6-963c-79c978013fa1 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 884.807700] env[61906]: DEBUG nova.network.neutron [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Updating instance_info_cache with network_info: [{"id": "58de857e-fc72-4f45-9da6-93623ceb23bc", "address": "fa:16:3e:34:df:8f", "network": {"id": "d123fdf4-6bb4-400b-a0bb-9d477672af2c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2099881656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d2198383d74dac90aa9727a62176df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58de857e-fc", "ovs_interfaceid": "58de857e-fc72-4f45-9da6-93623ceb23bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.829157] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 884.829157] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52b498e9-7a4b-b171-2530-ab211162ff9b" [ 884.829157] env[61906]: _type = "HttpNfcLease" [ 884.829157] env[61906]: } is ready. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 884.829470] env[61906]: DEBUG oslo_vmware.rw_handles [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 884.829470] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52b498e9-7a4b-b171-2530-ab211162ff9b" [ 884.829470] env[61906]: _type = "HttpNfcLease" [ 884.829470] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 884.830233] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6cbc731-b83d-45aa-bcb5-1f18bd84e36f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.841303] env[61906]: DEBUG nova.compute.manager [None req-067a1d58-ec9f-4fe7-bc35-ed5b951b5892 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 884.841303] env[61906]: DEBUG oslo_vmware.rw_handles [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5264888f-acba-b2b2-38e4-d1a6b31affde/disk-0.vmdk from lease info. {{(pid=61906) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 884.841303] env[61906]: DEBUG oslo_vmware.rw_handles [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5264888f-acba-b2b2-38e4-d1a6b31affde/disk-0.vmdk for reading. {{(pid=61906) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 884.842603] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a382b75e-39cf-41a7-8a10-afcbddebdaf4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.915839] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Acquiring lock "3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.916096] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Lock "3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.916293] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Acquiring lock "3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.916469] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Lock "3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.916628] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Lock "3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.918849] env[61906]: INFO nova.compute.manager [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Terminating instance [ 884.920450] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Acquiring lock "refresh_cache-3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.920596] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Acquired lock "refresh_cache-3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.920754] env[61906]: DEBUG nova.network.neutron [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 884.960511] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333236, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071078} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.960777] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 884.961534] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ddc178-f27d-4afb-8bb3-8d9c4ecac0f2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.984503] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 75b9738e-e9b6-435a-88bb-851982828d36/75b9738e-e9b6-435a-88bb-851982828d36.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 884.987786] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b45a0a4d-0c88-42e3-8d0e-6ab058d017b7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.008993] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333234, 'name': ReconfigVM_Task, 'duration_secs': 0.608051} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.010740] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Reconfigured VM instance instance-0000003e to attach disk [datastore1] 9f1253e6-4727-4dca-bad2-04f0c5424cdb/9f1253e6-4727-4dca-bad2-04f0c5424cdb.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 885.011112] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Waiting for the task: (returnval){ [ 885.011112] env[61906]: value = "task-1333238" [ 885.011112] env[61906]: _type = "Task" [ 885.011112] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.011334] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cc710438-5413-4c72-90cd-782325906278 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.023066] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333238, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.024480] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 885.024480] env[61906]: value = "task-1333239" [ 885.024480] env[61906]: _type = "Task" [ 885.024480] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.032350] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333239, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.064818] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1df06489-099a-435f-96e4-9aafe70b1403 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.106717] env[61906]: DEBUG nova.scheduler.client.report [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 885.120597] env[61906]: DEBUG oslo_vmware.api [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Task: {'id': task-1333237, 'name': PowerOffVM_Task, 'duration_secs': 0.479171} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.120753] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 885.120917] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 885.121174] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4e64d3b8-d77d-468c-b1c1-83ddf2fdf125 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.182150] env[61906]: DEBUG nova.compute.manager [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Received event network-vif-plugged-58de857e-fc72-4f45-9da6-93623ceb23bc {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 885.182473] env[61906]: DEBUG oslo_concurrency.lockutils [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] Acquiring lock "cc8870d0-eafb-4f73-aa79-b98f51370237-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.186017] env[61906]: DEBUG oslo_concurrency.lockutils [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] Lock "cc8870d0-eafb-4f73-aa79-b98f51370237-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.186017] env[61906]: DEBUG oslo_concurrency.lockutils [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] Lock "cc8870d0-eafb-4f73-aa79-b98f51370237-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.186017] env[61906]: DEBUG nova.compute.manager [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] No waiting events found dispatching network-vif-plugged-58de857e-fc72-4f45-9da6-93623ceb23bc {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 885.186017] env[61906]: WARNING nova.compute.manager [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Received unexpected event network-vif-plugged-58de857e-fc72-4f45-9da6-93623ceb23bc for instance with vm_state building and task_state spawning. [ 885.186017] env[61906]: DEBUG nova.compute.manager [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Received event network-changed-58de857e-fc72-4f45-9da6-93623ceb23bc {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 885.186017] env[61906]: DEBUG nova.compute.manager [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Refreshing instance network info cache due to event network-changed-58de857e-fc72-4f45-9da6-93623ceb23bc. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 885.186017] env[61906]: DEBUG oslo_concurrency.lockutils [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] Acquiring lock "refresh_cache-cc8870d0-eafb-4f73-aa79-b98f51370237" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.196090] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 885.196338] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 885.196521] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Deleting the datastore file [datastore2] 80a6a0b8-17c3-46e8-8861-6d1759a19111 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 885.196785] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-90ac4438-fbb2-4052-be87-63bfa9be181c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.203383] env[61906]: DEBUG oslo_vmware.api [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Waiting for the task: (returnval){ [ 885.203383] env[61906]: value = "task-1333241" [ 885.203383] env[61906]: _type = "Task" [ 885.203383] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.214324] env[61906]: DEBUG oslo_vmware.api [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Task: {'id': task-1333241, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.217501] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333231, 'name': CreateVM_Task, 'duration_secs': 1.596821} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.217661] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 885.218300] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.218457] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.218756] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 885.218990] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b68d2de-51de-424a-8212-923399844338 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.222865] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 885.222865] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]520a6a23-7f0d-d3c8-b096-b5317ff6d355" [ 885.222865] env[61906]: _type = "Task" [ 885.222865] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.230010] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]520a6a23-7f0d-d3c8-b096-b5317ff6d355, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.302119] env[61906]: DEBUG nova.compute.manager [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 885.309635] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Releasing lock "refresh_cache-cc8870d0-eafb-4f73-aa79-b98f51370237" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.310010] env[61906]: DEBUG nova.compute.manager [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Instance network_info: |[{"id": "58de857e-fc72-4f45-9da6-93623ceb23bc", "address": "fa:16:3e:34:df:8f", "network": {"id": "d123fdf4-6bb4-400b-a0bb-9d477672af2c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2099881656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d2198383d74dac90aa9727a62176df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58de857e-fc", "ovs_interfaceid": "58de857e-fc72-4f45-9da6-93623ceb23bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 885.310338] env[61906]: DEBUG oslo_concurrency.lockutils [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] Acquired lock "refresh_cache-cc8870d0-eafb-4f73-aa79-b98f51370237" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.310665] env[61906]: DEBUG nova.network.neutron [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Refreshing network info cache for port 58de857e-fc72-4f45-9da6-93623ceb23bc {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 885.312126] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:34:df:8f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cba18f15-a919-422e-a423-1e705e233389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '58de857e-fc72-4f45-9da6-93623ceb23bc', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 885.321816] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Creating folder: Project (f2d2198383d74dac90aa9727a62176df). Parent ref: group-v284713. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 885.329547] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-af579986-34e7-424e-9d9e-5126e7638522 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.340369] env[61906]: DEBUG nova.virt.hardware [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 885.340369] env[61906]: DEBUG nova.virt.hardware [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 885.340369] env[61906]: DEBUG nova.virt.hardware [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 885.340738] env[61906]: DEBUG nova.virt.hardware [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 885.341072] env[61906]: DEBUG nova.virt.hardware [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 885.341374] env[61906]: DEBUG nova.virt.hardware [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 885.341854] env[61906]: DEBUG nova.virt.hardware [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 885.342619] env[61906]: DEBUG nova.virt.hardware [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 885.342937] env[61906]: DEBUG nova.virt.hardware [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 885.343297] env[61906]: DEBUG nova.virt.hardware [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 885.343628] env[61906]: DEBUG nova.virt.hardware [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 885.345151] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ece9339-566e-4bea-8d3b-40440ebf5eaa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.349749] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Created folder: Project (f2d2198383d74dac90aa9727a62176df) in parent group-v284713. [ 885.350118] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Creating folder: Instances. Parent ref: group-v284765. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 885.350850] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-09d7244c-e8c7-4a60-be8a-fefd6632d185 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.356707] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a100a0cb-6f0d-4c46-8b30-2e2a5480b56e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.366022] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Created folder: Instances in parent group-v284765. [ 885.366022] env[61906]: DEBUG oslo.service.loopingcall [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 885.366022] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 885.366239] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5f377f74-41d6-42d3-bf71-59f41faa2315 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.396153] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 885.396153] env[61906]: value = "task-1333244" [ 885.396153] env[61906]: _type = "Task" [ 885.396153] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.405778] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333244, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.409766] env[61906]: INFO nova.compute.manager [None req-067a1d58-ec9f-4fe7-bc35-ed5b951b5892 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] instance snapshotting [ 885.410469] env[61906]: DEBUG nova.objects.instance [None req-067a1d58-ec9f-4fe7-bc35-ed5b951b5892 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Lazy-loading 'flavor' on Instance uuid 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.448266] env[61906]: DEBUG nova.network.neutron [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 885.527492] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333238, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.534875] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333239, 'name': Rename_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.549808] env[61906]: DEBUG nova.network.neutron [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.611477] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.333s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.611907] env[61906]: DEBUG nova.compute.manager [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 885.615791] env[61906]: DEBUG oslo_concurrency.lockutils [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.240s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.616921] env[61906]: DEBUG nova.objects.instance [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Lazy-loading 'resources' on Instance uuid 7e05fee8-842f-4e03-853a-58bac5a9ac0e {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.713609] env[61906]: DEBUG oslo_vmware.api [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Task: {'id': task-1333241, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.301619} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.714517] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 885.714517] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 885.714709] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 885.714925] env[61906]: INFO nova.compute.manager [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Took 1.13 seconds to destroy the instance on the hypervisor. [ 885.715296] env[61906]: DEBUG oslo.service.loopingcall [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 885.716069] env[61906]: DEBUG nova.compute.manager [-] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 885.716233] env[61906]: DEBUG nova.network.neutron [-] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 885.734562] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]520a6a23-7f0d-d3c8-b096-b5317ff6d355, 'name': SearchDatastore_Task, 'duration_secs': 0.017026} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.735018] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.735282] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 885.735661] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.735887] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.736180] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 885.736527] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-28f645c9-3f63-4a24-8d2d-ebc815b2215c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.745563] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 885.745867] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 885.746640] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3bbc0580-122d-432c-8b59-ab506f35f778 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.754735] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 885.754735] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52424704-fd10-cc5f-31d8-a3baa327fc62" [ 885.754735] env[61906]: _type = "Task" [ 885.754735] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.763193] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52424704-fd10-cc5f-31d8-a3baa327fc62, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.906984] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333244, 'name': CreateVM_Task, 'duration_secs': 0.442224} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.908051] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 885.908644] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.908773] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.909116] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 885.909386] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ea73c69-94d0-41bc-9369-303f13dfbf1b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.916334] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 885.916334] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]525dc5bf-8fbf-eff1-63c4-5f44250fa57f" [ 885.916334] env[61906]: _type = "Task" [ 885.916334] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.918078] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-079bc2c3-f9de-40fe-b27c-54e66e07dce6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.943826] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5da7ea5-3d1c-48a1-8186-939bc420581e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.947434] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]525dc5bf-8fbf-eff1-63c4-5f44250fa57f, 'name': SearchDatastore_Task, 'duration_secs': 0.011101} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.947846] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.948197] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 885.948431] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.987681] env[61906]: DEBUG nova.compute.manager [req-0657ccde-b392-487e-bf16-f53cf87c5a7f req-63beefbe-dbbd-4959-afd4-b7e106b64e63 service nova] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Received event network-vif-deleted-b6dc013b-172b-44b2-9f3c-a78cb9245964 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 885.988133] env[61906]: INFO nova.compute.manager [req-0657ccde-b392-487e-bf16-f53cf87c5a7f req-63beefbe-dbbd-4959-afd4-b7e106b64e63 service nova] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Neutron deleted interface b6dc013b-172b-44b2-9f3c-a78cb9245964; detaching it from the instance and deleting it from the info cache [ 885.988203] env[61906]: DEBUG nova.network.neutron [req-0657ccde-b392-487e-bf16-f53cf87c5a7f req-63beefbe-dbbd-4959-afd4-b7e106b64e63 service nova] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.024624] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333238, 'name': ReconfigVM_Task, 'duration_secs': 0.604993} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.026660] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 75b9738e-e9b6-435a-88bb-851982828d36/75b9738e-e9b6-435a-88bb-851982828d36.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 886.027903] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d9ccbe60-86b8-4403-a004-a872ed5f6728 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.040964] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Waiting for the task: (returnval){ [ 886.040964] env[61906]: value = "task-1333245" [ 886.040964] env[61906]: _type = "Task" [ 886.040964] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.040964] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333239, 'name': Rename_Task, 'duration_secs': 0.720815} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.041279] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 886.044344] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e7cd9b58-0876-446c-909a-8a81e59d492f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.053860] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Releasing lock "refresh_cache-3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.054406] env[61906]: DEBUG nova.compute.manager [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 886.054711] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 886.055083] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333245, 'name': Rename_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.057024] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7188e020-1004-46ca-83af-5ca6a7bab5a9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.060691] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 886.060691] env[61906]: value = "task-1333246" [ 886.060691] env[61906]: _type = "Task" [ 886.060691] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.066325] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 886.068135] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-452b0c79-6142-498e-9e99-129464210d00 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.073304] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333246, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.078682] env[61906]: DEBUG oslo_vmware.api [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Waiting for the task: (returnval){ [ 886.078682] env[61906]: value = "task-1333247" [ 886.078682] env[61906]: _type = "Task" [ 886.078682] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.091610] env[61906]: DEBUG oslo_vmware.api [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Task: {'id': task-1333247, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.123859] env[61906]: DEBUG nova.compute.utils [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 886.128296] env[61906]: DEBUG nova.compute.manager [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 886.128845] env[61906]: DEBUG nova.network.neutron [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 886.198083] env[61906]: DEBUG nova.policy [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c6c9ba4b5c814fa89c3cf8b74818e777', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '527d3de1bab74c4a8684edba72b6cb3f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 886.210742] env[61906]: DEBUG nova.network.neutron [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Updated VIF entry in instance network info cache for port 58de857e-fc72-4f45-9da6-93623ceb23bc. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 886.211325] env[61906]: DEBUG nova.network.neutron [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Updating instance_info_cache with network_info: [{"id": "58de857e-fc72-4f45-9da6-93623ceb23bc", "address": "fa:16:3e:34:df:8f", "network": {"id": "d123fdf4-6bb4-400b-a0bb-9d477672af2c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2099881656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d2198383d74dac90aa9727a62176df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58de857e-fc", "ovs_interfaceid": "58de857e-fc72-4f45-9da6-93623ceb23bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.268505] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52424704-fd10-cc5f-31d8-a3baa327fc62, 'name': SearchDatastore_Task, 'duration_secs': 0.0185} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.269663] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89cadc2f-4210-4ab8-86fc-dcbc179c55ae {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.286377] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 886.286377] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]520926dc-a670-abeb-5636-e48bd1142168" [ 886.286377] env[61906]: _type = "Task" [ 886.286377] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.296260] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]520926dc-a670-abeb-5636-e48bd1142168, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.382201] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d8e6c25-0aa1-48ea-ba7d-09a3d388a460 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.390567] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bc30ed3-caa6-49e7-ba52-762f6b057e79 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.426775] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-650c9cdd-9f6e-4f3d-b125-34a2d80c3228 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.435026] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be0626a-949d-490b-ad69-9c83aaf3ee7e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.450651] env[61906]: DEBUG nova.compute.provider_tree [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.459888] env[61906]: DEBUG nova.network.neutron [-] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.461437] env[61906]: DEBUG nova.compute.manager [None req-067a1d58-ec9f-4fe7-bc35-ed5b951b5892 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Instance disappeared during snapshot {{(pid=61906) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 886.465182] env[61906]: DEBUG nova.network.neutron [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Successfully updated port: 175a521b-1c70-41f6-963c-79c978013fa1 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 886.492821] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d43a44bd-29e1-4ebf-85ec-0eec6fb8a51d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.505487] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a592dc-fe6e-4cea-ba1e-47d3b86b437d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.540082] env[61906]: DEBUG nova.compute.manager [req-0657ccde-b392-487e-bf16-f53cf87c5a7f req-63beefbe-dbbd-4959-afd4-b7e106b64e63 service nova] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Detach interface failed, port_id=b6dc013b-172b-44b2-9f3c-a78cb9245964, reason: Instance 80a6a0b8-17c3-46e8-8861-6d1759a19111 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 886.550642] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333245, 'name': Rename_Task, 'duration_secs': 0.196365} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.550941] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 886.551217] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b3af826d-ec16-4993-8ba0-8e2c19979c14 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.561935] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Waiting for the task: (returnval){ [ 886.561935] env[61906]: value = "task-1333248" [ 886.561935] env[61906]: _type = "Task" [ 886.561935] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.576255] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333246, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.581808] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333248, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.596375] env[61906]: DEBUG oslo_vmware.api [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Task: {'id': task-1333247, 'name': PowerOffVM_Task, 'duration_secs': 0.116759} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.597324] env[61906]: DEBUG nova.network.neutron [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Successfully created port: f76ea64e-15c4-490e-9a5e-4c244195d983 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 886.599274] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 886.599458] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 886.600803] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1bb49079-3a5a-4677-8b54-b7256d7b43dd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.630284] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 886.630527] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 886.630718] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Deleting the datastore file [datastore1] 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 886.631012] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-082c7a99-a753-4eff-b2ff-ef074dd403f4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.636765] env[61906]: DEBUG nova.compute.manager [None req-067a1d58-ec9f-4fe7-bc35-ed5b951b5892 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Found 0 images (rotation: 2) {{(pid=61906) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 886.638126] env[61906]: DEBUG nova.compute.manager [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 886.646748] env[61906]: DEBUG oslo_vmware.api [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Waiting for the task: (returnval){ [ 886.646748] env[61906]: value = "task-1333250" [ 886.646748] env[61906]: _type = "Task" [ 886.646748] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.655710] env[61906]: DEBUG oslo_vmware.api [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Task: {'id': task-1333250, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.718447] env[61906]: DEBUG oslo_concurrency.lockutils [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] Releasing lock "refresh_cache-cc8870d0-eafb-4f73-aa79-b98f51370237" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.718804] env[61906]: DEBUG nova.compute.manager [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Received event network-changed-b6dc013b-172b-44b2-9f3c-a78cb9245964 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 886.718983] env[61906]: DEBUG nova.compute.manager [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Refreshing instance network info cache due to event network-changed-b6dc013b-172b-44b2-9f3c-a78cb9245964. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 886.719226] env[61906]: DEBUG oslo_concurrency.lockutils [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] Acquiring lock "refresh_cache-80a6a0b8-17c3-46e8-8861-6d1759a19111" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.719376] env[61906]: DEBUG oslo_concurrency.lockutils [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] Acquired lock "refresh_cache-80a6a0b8-17c3-46e8-8861-6d1759a19111" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.719614] env[61906]: DEBUG nova.network.neutron [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Refreshing network info cache for port b6dc013b-172b-44b2-9f3c-a78cb9245964 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 886.801681] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]520926dc-a670-abeb-5636-e48bd1142168, 'name': SearchDatastore_Task, 'duration_secs': 0.013887} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.802009] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.802290] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] cb34a0ce-d274-4167-8367-e19809cc9e4d/cb34a0ce-d274-4167-8367-e19809cc9e4d.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 886.802588] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.802778] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 886.802998] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-93f98c45-482d-4d1c-a3dc-0b9d483f000e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.805040] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fc5fc871-3dc4-4141-845c-5ee8719cc1ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.812097] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 886.812097] env[61906]: value = "task-1333251" [ 886.812097] env[61906]: _type = "Task" [ 886.812097] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.816417] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 886.816674] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 886.817776] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac79f0bc-b95d-4de5-9305-82dd8c33e165 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.822868] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333251, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.828831] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 886.828831] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52319624-5413-6e16-42c6-feca65dcb4e9" [ 886.828831] env[61906]: _type = "Task" [ 886.828831] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.834658] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52319624-5413-6e16-42c6-feca65dcb4e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.955031] env[61906]: DEBUG nova.scheduler.client.report [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.964912] env[61906]: INFO nova.compute.manager [-] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Took 1.25 seconds to deallocate network for instance. [ 886.969801] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "refresh_cache-7504dfa6-bedc-4701-b4fc-60e19e742276" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.970381] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquired lock "refresh_cache-7504dfa6-bedc-4701-b4fc-60e19e742276" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.970532] env[61906]: DEBUG nova.network.neutron [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 887.073932] env[61906]: DEBUG oslo_vmware.api [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333246, 'name': PowerOnVM_Task, 'duration_secs': 0.575094} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.074632] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 887.074849] env[61906]: DEBUG nova.compute.manager [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 887.075931] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70b5a9e0-230d-41a4-a542-ce8fb3296fe8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.081182] env[61906]: DEBUG oslo_vmware.api [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333248, 'name': PowerOnVM_Task, 'duration_secs': 0.484886} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.081794] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 887.082072] env[61906]: INFO nova.compute.manager [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Took 8.73 seconds to spawn the instance on the hypervisor. [ 887.082319] env[61906]: DEBUG nova.compute.manager [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 887.083162] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49cde489-4216-48b5-8a08-2e9f1ac4ca90 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.145201] env[61906]: INFO nova.virt.block_device [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Booting with volume 7f04fa3d-ec3b-4151-a46b-f232180344b9 at /dev/sda [ 887.161779] env[61906]: DEBUG oslo_vmware.api [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Task: {'id': task-1333250, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150273} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.162148] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 887.162346] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 887.162938] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 887.162938] env[61906]: INFO nova.compute.manager [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Took 1.11 seconds to destroy the instance on the hypervisor. [ 887.163122] env[61906]: DEBUG oslo.service.loopingcall [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 887.163287] env[61906]: DEBUG nova.compute.manager [-] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 887.163403] env[61906]: DEBUG nova.network.neutron [-] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 887.188676] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-24dc226e-953f-4336-8007-13920a554b7a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.198894] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fb7a81e-83f7-4248-8dac-da691bf6e256 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.229732] env[61906]: DEBUG nova.network.neutron [-] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 887.231117] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c1b7723a-65b3-434a-a787-ccf69ca71927 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.240030] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4662e73-a707-427c-af69-306cb789f35d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.251572] env[61906]: DEBUG nova.network.neutron [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 887.272387] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-515bb78c-09ff-4d29-b360-82bc38e3ca38 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.276452] env[61906]: DEBUG nova.compute.manager [req-8713acd6-7347-42b4-acdc-f29046b59403 req-bed8e9ef-c2e7-426f-9e5c-62778ba02b5b service nova] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Received event network-vif-plugged-175a521b-1c70-41f6-963c-79c978013fa1 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 887.276695] env[61906]: DEBUG oslo_concurrency.lockutils [req-8713acd6-7347-42b4-acdc-f29046b59403 req-bed8e9ef-c2e7-426f-9e5c-62778ba02b5b service nova] Acquiring lock "7504dfa6-bedc-4701-b4fc-60e19e742276-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.276900] env[61906]: DEBUG oslo_concurrency.lockutils [req-8713acd6-7347-42b4-acdc-f29046b59403 req-bed8e9ef-c2e7-426f-9e5c-62778ba02b5b service nova] Lock "7504dfa6-bedc-4701-b4fc-60e19e742276-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.277103] env[61906]: DEBUG oslo_concurrency.lockutils [req-8713acd6-7347-42b4-acdc-f29046b59403 req-bed8e9ef-c2e7-426f-9e5c-62778ba02b5b service nova] Lock "7504dfa6-bedc-4701-b4fc-60e19e742276-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.277247] env[61906]: DEBUG nova.compute.manager [req-8713acd6-7347-42b4-acdc-f29046b59403 req-bed8e9ef-c2e7-426f-9e5c-62778ba02b5b service nova] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] No waiting events found dispatching network-vif-plugged-175a521b-1c70-41f6-963c-79c978013fa1 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 887.277475] env[61906]: WARNING nova.compute.manager [req-8713acd6-7347-42b4-acdc-f29046b59403 req-bed8e9ef-c2e7-426f-9e5c-62778ba02b5b service nova] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Received unexpected event network-vif-plugged-175a521b-1c70-41f6-963c-79c978013fa1 for instance with vm_state building and task_state spawning. [ 887.277647] env[61906]: DEBUG nova.compute.manager [req-8713acd6-7347-42b4-acdc-f29046b59403 req-bed8e9ef-c2e7-426f-9e5c-62778ba02b5b service nova] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Received event network-changed-175a521b-1c70-41f6-963c-79c978013fa1 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 887.277809] env[61906]: DEBUG nova.compute.manager [req-8713acd6-7347-42b4-acdc-f29046b59403 req-bed8e9ef-c2e7-426f-9e5c-62778ba02b5b service nova] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Refreshing instance network info cache due to event network-changed-175a521b-1c70-41f6-963c-79c978013fa1. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 887.277980] env[61906]: DEBUG oslo_concurrency.lockutils [req-8713acd6-7347-42b4-acdc-f29046b59403 req-bed8e9ef-c2e7-426f-9e5c-62778ba02b5b service nova] Acquiring lock "refresh_cache-7504dfa6-bedc-4701-b4fc-60e19e742276" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.284220] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37eb8a28-de3e-4a31-b1fb-e3bdb3685d14 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.300104] env[61906]: DEBUG nova.virt.block_device [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Updating existing volume attachment record: d8b9bd27-d66d-4490-970e-10b4ef9bc353 {{(pid=61906) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 887.323375] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333251, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.336226] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52319624-5413-6e16-42c6-feca65dcb4e9, 'name': SearchDatastore_Task, 'duration_secs': 0.019271} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.337099] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e275c3c-e30c-43bb-96fb-afd971e1234d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.343084] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 887.343084] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]528df644-eed7-bb7c-02b6-73f9226755f5" [ 887.343084] env[61906]: _type = "Task" [ 887.343084] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.355472] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]528df644-eed7-bb7c-02b6-73f9226755f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.380969] env[61906]: DEBUG nova.network.neutron [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.461263] env[61906]: DEBUG oslo_concurrency.lockutils [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.845s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.463615] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 11.059s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.473450] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.487271] env[61906]: INFO nova.scheduler.client.report [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Deleted allocations for instance 7e05fee8-842f-4e03-853a-58bac5a9ac0e [ 887.516258] env[61906]: DEBUG nova.network.neutron [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 887.600591] env[61906]: DEBUG oslo_concurrency.lockutils [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.603121] env[61906]: INFO nova.compute.manager [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Took 26.06 seconds to build instance. [ 887.673433] env[61906]: DEBUG nova.network.neutron [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Updating instance_info_cache with network_info: [{"id": "175a521b-1c70-41f6-963c-79c978013fa1", "address": "fa:16:3e:dd:31:4c", "network": {"id": "c3bcef0f-44a7-44f7-8130-df77c0d3ac2d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-330384233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "129cdb06b47346b6895df45e73f7b2dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35e463c7-7d78-4d66-8efd-6127b1f3ee17", "external-id": "nsx-vlan-transportzone-175", "segmentation_id": 175, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap175a521b-1c", "ovs_interfaceid": "175a521b-1c70-41f6-963c-79c978013fa1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.733795] env[61906]: DEBUG nova.network.neutron [-] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.823882] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333251, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.736555} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.824172] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] cb34a0ce-d274-4167-8367-e19809cc9e4d/cb34a0ce-d274-4167-8367-e19809cc9e4d.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 887.824419] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 887.824676] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-901e6679-43c7-41af-9e82-12d427924b25 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.831185] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 887.831185] env[61906]: value = "task-1333252" [ 887.831185] env[61906]: _type = "Task" [ 887.831185] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.839651] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333252, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.853376] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]528df644-eed7-bb7c-02b6-73f9226755f5, 'name': SearchDatastore_Task, 'duration_secs': 0.066742} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.853691] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.853987] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] cc8870d0-eafb-4f73-aa79-b98f51370237/cc8870d0-eafb-4f73-aa79-b98f51370237.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 887.854292] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d5da0fff-02f7-4110-8040-0f608228272f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.861919] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 887.861919] env[61906]: value = "task-1333253" [ 887.861919] env[61906]: _type = "Task" [ 887.861919] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.870494] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333253, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.884486] env[61906]: DEBUG oslo_concurrency.lockutils [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] Releasing lock "refresh_cache-80a6a0b8-17c3-46e8-8861-6d1759a19111" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.884486] env[61906]: DEBUG nova.compute.manager [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Received event network-changed-3e04c5c7-8d7a-46ae-a2d7-52dc429650f9 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 887.884486] env[61906]: DEBUG nova.compute.manager [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Refreshing instance network info cache due to event network-changed-3e04c5c7-8d7a-46ae-a2d7-52dc429650f9. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 887.884659] env[61906]: DEBUG oslo_concurrency.lockutils [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] Acquiring lock "refresh_cache-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.884696] env[61906]: DEBUG oslo_concurrency.lockutils [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] Acquired lock "refresh_cache-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.885751] env[61906]: DEBUG nova.network.neutron [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Refreshing network info cache for port 3e04c5c7-8d7a-46ae-a2d7-52dc429650f9 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 887.995774] env[61906]: DEBUG oslo_concurrency.lockutils [None req-64d2705b-3695-46ca-a29d-51cd959574d6 tempest-ServersTestMultiNic-1233255236 tempest-ServersTestMultiNic-1233255236-project-member] Lock "7e05fee8-842f-4e03-853a-58bac5a9ac0e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.380s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.105112] env[61906]: DEBUG oslo_concurrency.lockutils [None req-550d20a0-0f7b-4ac4-a9ff-4b81bdd5ee6b tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Lock "75b9738e-e9b6-435a-88bb-851982828d36" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.684s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.175973] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Releasing lock "refresh_cache-7504dfa6-bedc-4701-b4fc-60e19e742276" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.176317] env[61906]: DEBUG nova.compute.manager [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Instance network_info: |[{"id": "175a521b-1c70-41f6-963c-79c978013fa1", "address": "fa:16:3e:dd:31:4c", "network": {"id": "c3bcef0f-44a7-44f7-8130-df77c0d3ac2d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-330384233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "129cdb06b47346b6895df45e73f7b2dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35e463c7-7d78-4d66-8efd-6127b1f3ee17", "external-id": "nsx-vlan-transportzone-175", "segmentation_id": 175, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap175a521b-1c", "ovs_interfaceid": "175a521b-1c70-41f6-963c-79c978013fa1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 888.177933] env[61906]: DEBUG oslo_concurrency.lockutils [req-8713acd6-7347-42b4-acdc-f29046b59403 req-bed8e9ef-c2e7-426f-9e5c-62778ba02b5b service nova] Acquired lock "refresh_cache-7504dfa6-bedc-4701-b4fc-60e19e742276" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.177933] env[61906]: DEBUG nova.network.neutron [req-8713acd6-7347-42b4-acdc-f29046b59403 req-bed8e9ef-c2e7-426f-9e5c-62778ba02b5b service nova] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Refreshing network info cache for port 175a521b-1c70-41f6-963c-79c978013fa1 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 888.178152] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dd:31:4c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35e463c7-7d78-4d66-8efd-6127b1f3ee17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '175a521b-1c70-41f6-963c-79c978013fa1', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 888.186870] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Creating folder: Project (129cdb06b47346b6895df45e73f7b2dd). Parent ref: group-v284713. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 888.188067] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6e86b1f2-ac10-444e-80a5-78eb04dfddcb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.201973] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Created folder: Project (129cdb06b47346b6895df45e73f7b2dd) in parent group-v284713. [ 888.201973] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Creating folder: Instances. Parent ref: group-v284768. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 888.201973] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4ba1b907-269c-4782-9bf3-0b0738e261a3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.210456] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Created folder: Instances in parent group-v284768. [ 888.210962] env[61906]: DEBUG oslo.service.loopingcall [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 888.211337] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 888.211875] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-95267ca0-359e-4a28-9a76-a988f1ff714e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.233481] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 888.233481] env[61906]: value = "task-1333256" [ 888.233481] env[61906]: _type = "Task" [ 888.233481] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.237009] env[61906]: INFO nova.compute.manager [-] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Took 1.07 seconds to deallocate network for instance. [ 888.250166] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333256, 'name': CreateVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.342977] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333252, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070811} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.343473] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 888.344474] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-531e5d9f-a8cf-4639-aa4a-1d0b82c2af1c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.375279] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] cb34a0ce-d274-4167-8367-e19809cc9e4d/cb34a0ce-d274-4167-8367-e19809cc9e4d.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 888.378992] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2f3eefe1-5e97-43c6-974d-5e34460cc0ac {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.398427] env[61906]: DEBUG oslo_concurrency.lockutils [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "9f1253e6-4727-4dca-bad2-04f0c5424cdb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.398427] env[61906]: DEBUG oslo_concurrency.lockutils [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "9f1253e6-4727-4dca-bad2-04f0c5424cdb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.398427] env[61906]: DEBUG oslo_concurrency.lockutils [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "9f1253e6-4727-4dca-bad2-04f0c5424cdb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.398427] env[61906]: DEBUG oslo_concurrency.lockutils [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "9f1253e6-4727-4dca-bad2-04f0c5424cdb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.398427] env[61906]: DEBUG oslo_concurrency.lockutils [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "9f1253e6-4727-4dca-bad2-04f0c5424cdb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.400273] env[61906]: INFO nova.compute.manager [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Terminating instance [ 888.402599] env[61906]: DEBUG nova.compute.manager [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 888.402843] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 888.408381] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5baa948-0c68-437b-b434-545b86e36485 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.419039] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333253, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.425709] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 888.425864] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 888.425864] env[61906]: value = "task-1333257" [ 888.425864] env[61906]: _type = "Task" [ 888.425864] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.428023] env[61906]: DEBUG nova.compute.manager [req-2d717047-7480-4452-bfb3-00edeb2f09c2 req-ffc2c6bc-14e7-4e99-ab25-32dd8689093e service nova] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Received event network-vif-plugged-f76ea64e-15c4-490e-9a5e-4c244195d983 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.428023] env[61906]: DEBUG oslo_concurrency.lockutils [req-2d717047-7480-4452-bfb3-00edeb2f09c2 req-ffc2c6bc-14e7-4e99-ab25-32dd8689093e service nova] Acquiring lock "da493512-d996-4de7-9f47-cadcbc4fbcb9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.428023] env[61906]: DEBUG oslo_concurrency.lockutils [req-2d717047-7480-4452-bfb3-00edeb2f09c2 req-ffc2c6bc-14e7-4e99-ab25-32dd8689093e service nova] Lock "da493512-d996-4de7-9f47-cadcbc4fbcb9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.428409] env[61906]: DEBUG oslo_concurrency.lockutils [req-2d717047-7480-4452-bfb3-00edeb2f09c2 req-ffc2c6bc-14e7-4e99-ab25-32dd8689093e service nova] Lock "da493512-d996-4de7-9f47-cadcbc4fbcb9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.428409] env[61906]: DEBUG nova.compute.manager [req-2d717047-7480-4452-bfb3-00edeb2f09c2 req-ffc2c6bc-14e7-4e99-ab25-32dd8689093e service nova] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] No waiting events found dispatching network-vif-plugged-f76ea64e-15c4-490e-9a5e-4c244195d983 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 888.428533] env[61906]: WARNING nova.compute.manager [req-2d717047-7480-4452-bfb3-00edeb2f09c2 req-ffc2c6bc-14e7-4e99-ab25-32dd8689093e service nova] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Received unexpected event network-vif-plugged-f76ea64e-15c4-490e-9a5e-4c244195d983 for instance with vm_state building and task_state block_device_mapping. [ 888.429609] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-06553635-17c2-4a78-bdff-01f9d3ed5ef5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.448614] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333257, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.448614] env[61906]: DEBUG oslo_vmware.api [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 888.448614] env[61906]: value = "task-1333258" [ 888.448614] env[61906]: _type = "Task" [ 888.448614] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.460881] env[61906]: DEBUG oslo_vmware.api [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333258, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.497930] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 9f1253e6-4727-4dca-bad2-04f0c5424cdb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.498412] env[61906]: WARNING nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 80a6a0b8-17c3-46e8-8861-6d1759a19111 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 888.498412] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance d60575a6-af13-4d81-95d1-081d0f3e2ce6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.498412] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.498646] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.498646] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 75b9738e-e9b6-435a-88bb-851982828d36 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.498933] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance cb34a0ce-d274-4167-8367-e19809cc9e4d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.498933] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance cc8870d0-eafb-4f73-aa79-b98f51370237 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.499080] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 7504dfa6-bedc-4701-b4fc-60e19e742276 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.499599] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance da493512-d996-4de7-9f47-cadcbc4fbcb9 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.521750] env[61906]: INFO nova.compute.manager [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Rescuing [ 888.522483] env[61906]: DEBUG oslo_concurrency.lockutils [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Acquiring lock "refresh_cache-75b9738e-e9b6-435a-88bb-851982828d36" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.522617] env[61906]: DEBUG oslo_concurrency.lockutils [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Acquired lock "refresh_cache-75b9738e-e9b6-435a-88bb-851982828d36" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.522924] env[61906]: DEBUG nova.network.neutron [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 888.536633] env[61906]: DEBUG nova.network.neutron [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Successfully updated port: f76ea64e-15c4-490e-9a5e-4c244195d983 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 888.743602] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.748913] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333256, 'name': CreateVM_Task, 'duration_secs': 0.417717} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.749063] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 888.749819] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.749986] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.750386] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 888.750679] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-caf073f3-b6d5-4b39-9a99-914fafc0532a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.756210] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 888.756210] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5213136e-3777-3a54-6eb6-694990c5f3f1" [ 888.756210] env[61906]: _type = "Task" [ 888.756210] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.765546] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5213136e-3777-3a54-6eb6-694990c5f3f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.802562] env[61906]: DEBUG nova.network.neutron [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Updated VIF entry in instance network info cache for port 3e04c5c7-8d7a-46ae-a2d7-52dc429650f9. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 888.803164] env[61906]: DEBUG nova.network.neutron [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Updating instance_info_cache with network_info: [{"id": "3e04c5c7-8d7a-46ae-a2d7-52dc429650f9", "address": "fa:16:3e:f1:15:76", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e04c5c7-8d", "ovs_interfaceid": "3e04c5c7-8d7a-46ae-a2d7-52dc429650f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.881429] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333253, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.828655} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.884399] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] cc8870d0-eafb-4f73-aa79-b98f51370237/cc8870d0-eafb-4f73-aa79-b98f51370237.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 888.884399] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 888.884399] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ad2e299e-57fa-40ae-b150-2900b3a27121 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.890266] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 888.890266] env[61906]: value = "task-1333259" [ 888.890266] env[61906]: _type = "Task" [ 888.890266] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.903103] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333259, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.913939] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.914222] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.945027] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333257, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.961522] env[61906]: DEBUG oslo_vmware.api [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333258, 'name': PowerOffVM_Task, 'duration_secs': 0.399471} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.965530] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 888.965618] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 888.969355] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-09f6dce4-3085-4a77-b808-b540003e5633 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.003700] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance b1159533-c970-49d6-af42-b954b20d92fb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 889.038104] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 889.038421] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 889.038661] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Deleting the datastore file [datastore1] 9f1253e6-4727-4dca-bad2-04f0c5424cdb {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 889.039325] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Acquiring lock "refresh_cache-da493512-d996-4de7-9f47-cadcbc4fbcb9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.039522] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Acquired lock "refresh_cache-da493512-d996-4de7-9f47-cadcbc4fbcb9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.039731] env[61906]: DEBUG nova.network.neutron [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 889.041601] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d5caa69c-516f-4b3f-bbfd-a119c1d265c4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.049487] env[61906]: DEBUG oslo_vmware.api [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 889.049487] env[61906]: value = "task-1333261" [ 889.049487] env[61906]: _type = "Task" [ 889.049487] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.053701] env[61906]: DEBUG nova.network.neutron [req-8713acd6-7347-42b4-acdc-f29046b59403 req-bed8e9ef-c2e7-426f-9e5c-62778ba02b5b service nova] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Updated VIF entry in instance network info cache for port 175a521b-1c70-41f6-963c-79c978013fa1. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 889.054073] env[61906]: DEBUG nova.network.neutron [req-8713acd6-7347-42b4-acdc-f29046b59403 req-bed8e9ef-c2e7-426f-9e5c-62778ba02b5b service nova] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Updating instance_info_cache with network_info: [{"id": "175a521b-1c70-41f6-963c-79c978013fa1", "address": "fa:16:3e:dd:31:4c", "network": {"id": "c3bcef0f-44a7-44f7-8130-df77c0d3ac2d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-330384233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "129cdb06b47346b6895df45e73f7b2dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35e463c7-7d78-4d66-8efd-6127b1f3ee17", "external-id": "nsx-vlan-transportzone-175", "segmentation_id": 175, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap175a521b-1c", "ovs_interfaceid": "175a521b-1c70-41f6-963c-79c978013fa1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.062319] env[61906]: DEBUG oslo_vmware.api [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333261, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.244266] env[61906]: DEBUG nova.network.neutron [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Updating instance_info_cache with network_info: [{"id": "8bcec3ac-5104-4e72-af50-08fae9dc156c", "address": "fa:16:3e:36:f8:fd", "network": {"id": "eaf64f00-dd3b-41c3-b679-3ee2f53735a0", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-255429219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "5375ff2ae28b4427a3c4541ee3d2729a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20fbc86c-0935-406c-9659-3ffc33fc0d08", "external-id": "nsx-vlan-transportzone-191", "segmentation_id": 191, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bcec3ac-51", "ovs_interfaceid": "8bcec3ac-5104-4e72-af50-08fae9dc156c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.267049] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5213136e-3777-3a54-6eb6-694990c5f3f1, 'name': SearchDatastore_Task, 'duration_secs': 0.014617} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.267536] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.267639] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 889.267833] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.267983] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.268186] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 889.268461] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d199a862-78a1-43fc-ae2d-6dc53e36152a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.280640] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 889.280856] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 889.281567] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0bc17856-7f44-4570-b895-a3ae9e08efa7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.287071] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 889.287071] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52a60e3c-e7ea-d171-6cf8-d6461b21cf03" [ 889.287071] env[61906]: _type = "Task" [ 889.287071] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.295089] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52a60e3c-e7ea-d171-6cf8-d6461b21cf03, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.306725] env[61906]: DEBUG oslo_concurrency.lockutils [req-489ce156-bcd3-42f9-a1cd-9eed5680404e req-94a06c77-3324-4821-9c21-d6877709afaf service nova] Releasing lock "refresh_cache-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.401867] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333259, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.099809} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.402267] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 889.403202] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8114f1bf-e754-4883-8f03-dee7c423086a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.418741] env[61906]: DEBUG nova.compute.manager [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 889.430571] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] cc8870d0-eafb-4f73-aa79-b98f51370237/cc8870d0-eafb-4f73-aa79-b98f51370237.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 889.431506] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e91a0c4-1914-4552-aa3f-3a4212ccaf33 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.458064] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333257, 'name': ReconfigVM_Task, 'duration_secs': 0.577937} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.459630] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Reconfigured VM instance instance-00000044 to attach disk [datastore1] cb34a0ce-d274-4167-8367-e19809cc9e4d/cb34a0ce-d274-4167-8367-e19809cc9e4d.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 889.460400] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 889.460400] env[61906]: value = "task-1333262" [ 889.460400] env[61906]: _type = "Task" [ 889.460400] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.460626] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c05653e0-6e3e-4611-bda4-8e2d31004008 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.469365] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 889.469365] env[61906]: value = "task-1333263" [ 889.469365] env[61906]: _type = "Task" [ 889.469365] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.474275] env[61906]: DEBUG nova.compute.manager [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 889.474856] env[61906]: DEBUG nova.virt.hardware [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 889.475140] env[61906]: DEBUG nova.virt.hardware [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 889.475342] env[61906]: DEBUG nova.virt.hardware [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 889.475556] env[61906]: DEBUG nova.virt.hardware [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 889.476313] env[61906]: DEBUG nova.virt.hardware [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 889.476460] env[61906]: DEBUG nova.virt.hardware [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 889.476735] env[61906]: DEBUG nova.virt.hardware [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 889.476959] env[61906]: DEBUG nova.virt.hardware [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 889.477167] env[61906]: DEBUG nova.virt.hardware [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 889.477345] env[61906]: DEBUG nova.virt.hardware [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 889.477570] env[61906]: DEBUG nova.virt.hardware [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 889.477900] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333262, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.483499] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a63bed8f-b5d2-4c2c-8164-e58a66db0bf2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.495400] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333263, 'name': Rename_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.501236] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3362d058-9dc6-4090-af78-c13c4181b771 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.509398] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 889.557709] env[61906]: DEBUG oslo_concurrency.lockutils [req-8713acd6-7347-42b4-acdc-f29046b59403 req-bed8e9ef-c2e7-426f-9e5c-62778ba02b5b service nova] Releasing lock "refresh_cache-7504dfa6-bedc-4701-b4fc-60e19e742276" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.562118] env[61906]: DEBUG oslo_vmware.api [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333261, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.410966} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.562541] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 889.562874] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 889.563224] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 889.563512] env[61906]: INFO nova.compute.manager [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Took 1.16 seconds to destroy the instance on the hypervisor. [ 889.563785] env[61906]: DEBUG oslo.service.loopingcall [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 889.564050] env[61906]: DEBUG nova.compute.manager [-] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 889.564158] env[61906]: DEBUG nova.network.neutron [-] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 889.587038] env[61906]: DEBUG nova.network.neutron [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 889.747386] env[61906]: DEBUG oslo_concurrency.lockutils [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Releasing lock "refresh_cache-75b9738e-e9b6-435a-88bb-851982828d36" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.802968] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52a60e3c-e7ea-d171-6cf8-d6461b21cf03, 'name': SearchDatastore_Task, 'duration_secs': 0.041104} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.803924] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2049523-43ec-4787-acb3-d9ff1a849447 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.812634] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 889.812634] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5210723b-3626-fc59-9642-26185baa6fb1" [ 889.812634] env[61906]: _type = "Task" [ 889.812634] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.816445] env[61906]: DEBUG nova.network.neutron [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Updating instance_info_cache with network_info: [{"id": "f76ea64e-15c4-490e-9a5e-4c244195d983", "address": "fa:16:3e:68:ee:08", "network": {"id": "b5a4d04c-7682-4981-9e34-c416774c62a0", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-554429652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "527d3de1bab74c4a8684edba72b6cb3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c979f78-8597-41f8-b1de-995014032689", "external-id": "nsx-vlan-transportzone-477", "segmentation_id": 477, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf76ea64e-15", "ovs_interfaceid": "f76ea64e-15c4-490e-9a5e-4c244195d983", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.823031] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5210723b-3626-fc59-9642-26185baa6fb1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.953798] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.973839] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333262, 'name': ReconfigVM_Task, 'duration_secs': 0.484165} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.974351] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Reconfigured VM instance instance-00000045 to attach disk [datastore1] cc8870d0-eafb-4f73-aa79-b98f51370237/cc8870d0-eafb-4f73-aa79-b98f51370237.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 889.978863] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5fe32cb0-9a82-44aa-be6b-cb7cecd41017 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.987057] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333263, 'name': Rename_Task, 'duration_secs': 0.201426} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.991718] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 889.991718] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 889.991718] env[61906]: value = "task-1333264" [ 889.991718] env[61906]: _type = "Task" [ 889.991718] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.991718] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5eb77162-8800-4731-9d9a-ed619c1614f9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.005042] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333264, 'name': Rename_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.008603] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 890.008603] env[61906]: value = "task-1333265" [ 890.008603] env[61906]: _type = "Task" [ 890.008603] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.019207] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 0e86bfdf-9002-4a2d-a2ac-af35921b020e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 890.020752] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333265, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.280965] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 890.280965] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a20e029b-a5ae-4507-92cf-20afc2a18f8e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.287141] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Waiting for the task: (returnval){ [ 890.287141] env[61906]: value = "task-1333266" [ 890.287141] env[61906]: _type = "Task" [ 890.287141] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.296879] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333266, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.321703] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Releasing lock "refresh_cache-da493512-d996-4de7-9f47-cadcbc4fbcb9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.322128] env[61906]: DEBUG nova.compute.manager [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Instance network_info: |[{"id": "f76ea64e-15c4-490e-9a5e-4c244195d983", "address": "fa:16:3e:68:ee:08", "network": {"id": "b5a4d04c-7682-4981-9e34-c416774c62a0", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-554429652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "527d3de1bab74c4a8684edba72b6cb3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c979f78-8597-41f8-b1de-995014032689", "external-id": "nsx-vlan-transportzone-477", "segmentation_id": 477, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf76ea64e-15", "ovs_interfaceid": "f76ea64e-15c4-490e-9a5e-4c244195d983", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 890.322525] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5210723b-3626-fc59-9642-26185baa6fb1, 'name': SearchDatastore_Task, 'duration_secs': 0.019465} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.322936] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:68:ee:08', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8c979f78-8597-41f8-b1de-995014032689', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f76ea64e-15c4-490e-9a5e-4c244195d983', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 890.331454] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Creating folder: Project (527d3de1bab74c4a8684edba72b6cb3f). Parent ref: group-v284713. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 890.331628] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.331885] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 7504dfa6-bedc-4701-b4fc-60e19e742276/7504dfa6-bedc-4701-b4fc-60e19e742276.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 890.332274] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-afb208c9-2f82-46f2-b2c2-cf9e184ccaa0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.335054] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1462f45c-1c60-40cc-866a-161d603b3fdb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.343318] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 890.343318] env[61906]: value = "task-1333268" [ 890.343318] env[61906]: _type = "Task" [ 890.343318] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.348620] env[61906]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 890.348849] env[61906]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61906) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 890.349263] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Folder already exists: Project (527d3de1bab74c4a8684edba72b6cb3f). Parent ref: group-v284713. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 890.349470] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Creating folder: Instances. Parent ref: group-v284733. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 890.350250] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8cc5756b-5c7a-46cf-8889-cdfe8975ac79 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.356619] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333268, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.366510] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Created folder: Instances in parent group-v284733. [ 890.366978] env[61906]: DEBUG oslo.service.loopingcall [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.367398] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 890.367621] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7528891c-6f0e-42e4-9ecc-22b4c6b4f44b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.393120] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 890.393120] env[61906]: value = "task-1333270" [ 890.393120] env[61906]: _type = "Task" [ 890.393120] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.404717] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333270, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.455882] env[61906]: DEBUG nova.compute.manager [req-c6ea175b-5c7d-44eb-b637-cf9642ee691b req-0f4a1f55-4a5c-4792-81d5-9c852962ffd8 service nova] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Received event network-changed-f76ea64e-15c4-490e-9a5e-4c244195d983 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 890.456246] env[61906]: DEBUG nova.compute.manager [req-c6ea175b-5c7d-44eb-b637-cf9642ee691b req-0f4a1f55-4a5c-4792-81d5-9c852962ffd8 service nova] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Refreshing instance network info cache due to event network-changed-f76ea64e-15c4-490e-9a5e-4c244195d983. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 890.456549] env[61906]: DEBUG oslo_concurrency.lockutils [req-c6ea175b-5c7d-44eb-b637-cf9642ee691b req-0f4a1f55-4a5c-4792-81d5-9c852962ffd8 service nova] Acquiring lock "refresh_cache-da493512-d996-4de7-9f47-cadcbc4fbcb9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.456733] env[61906]: DEBUG oslo_concurrency.lockutils [req-c6ea175b-5c7d-44eb-b637-cf9642ee691b req-0f4a1f55-4a5c-4792-81d5-9c852962ffd8 service nova] Acquired lock "refresh_cache-da493512-d996-4de7-9f47-cadcbc4fbcb9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.456901] env[61906]: DEBUG nova.network.neutron [req-c6ea175b-5c7d-44eb-b637-cf9642ee691b req-0f4a1f55-4a5c-4792-81d5-9c852962ffd8 service nova] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Refreshing network info cache for port f76ea64e-15c4-490e-9a5e-4c244195d983 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 890.492664] env[61906]: DEBUG nova.network.neutron [-] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.507337] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333264, 'name': Rename_Task, 'duration_secs': 0.233062} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.507731] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 890.507731] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-06b74886-b084-4515-a403-67af8b7dcda5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.522778] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 890.523117] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 890.523287] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2240MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 890.526224] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333265, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.529304] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 890.529304] env[61906]: value = "task-1333271" [ 890.529304] env[61906]: _type = "Task" [ 890.529304] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.540101] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333271, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.798512] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333266, 'name': PowerOffVM_Task, 'duration_secs': 0.267015} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.800574] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 890.801703] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c606635-684a-4864-8d64-9ce9471c6281 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.805939] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83b44314-9489-4ad4-8559-7a8a8002448f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.832517] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e314406-0aee-4f3b-b2cc-31ecff43e092 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.837464] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b2cec95-546a-48ba-b886-e9a8356aae4d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.886854] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc8c7eb4-1e7e-4b5f-8041-f3dd8d1c7d7b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.896178] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333268, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.903791] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcce81d8-30db-4d53-8169-259147b9ca38 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.916914] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333270, 'name': CreateVM_Task} progress is 25%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.929274] env[61906]: DEBUG nova.compute.provider_tree [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.930895] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 890.931366] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-37e1f21d-cc0e-4019-9dee-9df41a1643a5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.939866] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Waiting for the task: (returnval){ [ 890.939866] env[61906]: value = "task-1333272" [ 890.939866] env[61906]: _type = "Task" [ 890.939866] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.950441] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] VM already powered off {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 890.950799] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 890.950961] env[61906]: DEBUG oslo_concurrency.lockutils [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.951169] env[61906]: DEBUG oslo_concurrency.lockutils [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.951417] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 890.951660] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6e1bda6a-e48e-47c7-872e-df399ea3986c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.970239] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 890.970239] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 890.970239] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-274a3bc2-956d-4fa2-b9b7-e4e6e10fb649 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.981379] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Waiting for the task: (returnval){ [ 890.981379] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52471fe6-eafd-27de-60dd-6ec12aafde2e" [ 890.981379] env[61906]: _type = "Task" [ 890.981379] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.992962] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52471fe6-eafd-27de-60dd-6ec12aafde2e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.996674] env[61906]: INFO nova.compute.manager [-] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Took 1.43 seconds to deallocate network for instance. [ 891.020190] env[61906]: DEBUG oslo_vmware.api [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333265, 'name': PowerOnVM_Task, 'duration_secs': 0.721728} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.020520] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 891.020725] env[61906]: INFO nova.compute.manager [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Took 10.33 seconds to spawn the instance on the hypervisor. [ 891.020932] env[61906]: DEBUG nova.compute.manager [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 891.021824] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c3ac858-99b8-4424-8341-2afce3dc170f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.040735] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333271, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.399087] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333268, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.679518} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.402817] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 7504dfa6-bedc-4701-b4fc-60e19e742276/7504dfa6-bedc-4701-b4fc-60e19e742276.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 891.403160] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 891.403496] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4b2ba6c6-d039-42dc-8e55-4b6247e68cf7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.412804] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333270, 'name': CreateVM_Task, 'duration_secs': 0.74588} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.414242] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 891.414657] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 891.414657] env[61906]: value = "task-1333273" [ 891.414657] env[61906]: _type = "Task" [ 891.414657] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.415412] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'device_type': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284737', 'volume_id': '7f04fa3d-ec3b-4151-a46b-f232180344b9', 'name': 'volume-7f04fa3d-ec3b-4151-a46b-f232180344b9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'da493512-d996-4de7-9f47-cadcbc4fbcb9', 'attached_at': '', 'detached_at': '', 'volume_id': '7f04fa3d-ec3b-4151-a46b-f232180344b9', 'serial': '7f04fa3d-ec3b-4151-a46b-f232180344b9'}, 'attachment_id': 'd8b9bd27-d66d-4490-970e-10b4ef9bc353', 'delete_on_termination': True, 'boot_index': 0, 'disk_bus': None, 'guest_format': None, 'mount_device': '/dev/sda', 'volume_type': None}], 'swap': None} {{(pid=61906) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 891.415695] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Root volume attach. Driver type: vmdk {{(pid=61906) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 891.416568] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7dbf124-b70c-4c77-bb9f-70e95d850496 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.431569] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333273, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.433819] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1256ea6f-3aff-4da6-8a24-b2fb344385c5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.435782] env[61906]: DEBUG nova.scheduler.client.report [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 891.443257] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aed69cef-9e16-4d59-adf1-25a67882a84a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.449992] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-ab0a4bc7-1a5a-443d-a170-5e692606fd6a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.458552] env[61906]: DEBUG oslo_vmware.api [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Waiting for the task: (returnval){ [ 891.458552] env[61906]: value = "task-1333274" [ 891.458552] env[61906]: _type = "Task" [ 891.458552] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.466966] env[61906]: DEBUG oslo_vmware.api [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Task: {'id': task-1333274, 'name': RelocateVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.494638] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52471fe6-eafd-27de-60dd-6ec12aafde2e, 'name': SearchDatastore_Task, 'duration_secs': 0.022146} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.495446] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44fb5800-9e09-438b-a5c7-95c2bb1a07da {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.501540] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Waiting for the task: (returnval){ [ 891.501540] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]524dd54c-c9c8-1638-5411-939ad564a98c" [ 891.501540] env[61906]: _type = "Task" [ 891.501540] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.505700] env[61906]: DEBUG oslo_concurrency.lockutils [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.511213] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]524dd54c-c9c8-1638-5411-939ad564a98c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.557562] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333271, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.557562] env[61906]: INFO nova.compute.manager [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Took 28.11 seconds to build instance. [ 891.655255] env[61906]: DEBUG nova.network.neutron [req-c6ea175b-5c7d-44eb-b637-cf9642ee691b req-0f4a1f55-4a5c-4792-81d5-9c852962ffd8 service nova] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Updated VIF entry in instance network info cache for port f76ea64e-15c4-490e-9a5e-4c244195d983. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 891.655674] env[61906]: DEBUG nova.network.neutron [req-c6ea175b-5c7d-44eb-b637-cf9642ee691b req-0f4a1f55-4a5c-4792-81d5-9c852962ffd8 service nova] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Updating instance_info_cache with network_info: [{"id": "f76ea64e-15c4-490e-9a5e-4c244195d983", "address": "fa:16:3e:68:ee:08", "network": {"id": "b5a4d04c-7682-4981-9e34-c416774c62a0", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-554429652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "527d3de1bab74c4a8684edba72b6cb3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c979f78-8597-41f8-b1de-995014032689", "external-id": "nsx-vlan-transportzone-477", "segmentation_id": 477, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf76ea64e-15", "ovs_interfaceid": "f76ea64e-15c4-490e-9a5e-4c244195d983", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.927386] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333273, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.156337} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.927776] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 891.928752] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f58f5018-dbff-409d-8696-7ee2a68b0468 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.945228] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 891.945421] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.482s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.959038] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 7504dfa6-bedc-4701-b4fc-60e19e742276/7504dfa6-bedc-4701-b4fc-60e19e742276.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 891.959489] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.438s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.961174] env[61906]: INFO nova.compute.claims [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 891.964142] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9767fc4f-d09e-4f71-8598-17905debc261 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.991906] env[61906]: DEBUG oslo_vmware.api [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Task: {'id': task-1333274, 'name': RelocateVM_Task} progress is 20%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.991906] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 891.991906] env[61906]: value = "task-1333275" [ 891.991906] env[61906]: _type = "Task" [ 891.991906] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.002531] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333275, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.011314] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]524dd54c-c9c8-1638-5411-939ad564a98c, 'name': SearchDatastore_Task, 'duration_secs': 0.012721} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.011620] env[61906]: DEBUG oslo_concurrency.lockutils [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.011897] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 75b9738e-e9b6-435a-88bb-851982828d36/a9502563-1f96-46a5-b87e-12c0e9375e6e-rescue.vmdk. {{(pid=61906) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 892.012247] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b5755116-5537-4788-9df7-7f26ddd45774 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.018620] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Waiting for the task: (returnval){ [ 892.018620] env[61906]: value = "task-1333276" [ 892.018620] env[61906]: _type = "Task" [ 892.018620] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.028762] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333276, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.047023] env[61906]: DEBUG oslo_vmware.api [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333271, 'name': PowerOnVM_Task, 'duration_secs': 1.493644} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.047023] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 892.047023] env[61906]: INFO nova.compute.manager [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Took 9.02 seconds to spawn the instance on the hypervisor. [ 892.047023] env[61906]: DEBUG nova.compute.manager [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 892.047023] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3af0ffec-cc6b-4642-a7cf-e048d9a452c2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.062089] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ddd502eb-b1ec-4041-932a-7e9ace325c11 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "cb34a0ce-d274-4167-8367-e19809cc9e4d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 87.398s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.158114] env[61906]: DEBUG oslo_concurrency.lockutils [req-c6ea175b-5c7d-44eb-b637-cf9642ee691b req-0f4a1f55-4a5c-4792-81d5-9c852962ffd8 service nova] Releasing lock "refresh_cache-da493512-d996-4de7-9f47-cadcbc4fbcb9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.158515] env[61906]: DEBUG nova.compute.manager [req-c6ea175b-5c7d-44eb-b637-cf9642ee691b req-0f4a1f55-4a5c-4792-81d5-9c852962ffd8 service nova] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Received event network-vif-deleted-61a42623-e2e5-4245-9560-be57af9a8397 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.158716] env[61906]: INFO nova.compute.manager [req-c6ea175b-5c7d-44eb-b637-cf9642ee691b req-0f4a1f55-4a5c-4792-81d5-9c852962ffd8 service nova] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Neutron deleted interface 61a42623-e2e5-4245-9560-be57af9a8397; detaching it from the instance and deleting it from the info cache [ 892.158898] env[61906]: DEBUG nova.network.neutron [req-c6ea175b-5c7d-44eb-b637-cf9642ee691b req-0f4a1f55-4a5c-4792-81d5-9c852962ffd8 service nova] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.477817] env[61906]: DEBUG oslo_vmware.api [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Task: {'id': task-1333274, 'name': RelocateVM_Task, 'duration_secs': 0.594485} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.478179] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Volume attach. Driver type: vmdk {{(pid=61906) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 892.478395] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284737', 'volume_id': '7f04fa3d-ec3b-4151-a46b-f232180344b9', 'name': 'volume-7f04fa3d-ec3b-4151-a46b-f232180344b9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'da493512-d996-4de7-9f47-cadcbc4fbcb9', 'attached_at': '', 'detached_at': '', 'volume_id': '7f04fa3d-ec3b-4151-a46b-f232180344b9', 'serial': '7f04fa3d-ec3b-4151-a46b-f232180344b9'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 892.479263] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f429353f-6237-4fd7-84c4-da0946bd7750 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.507752] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08174662-3cb5-45aa-b3f9-0cbed7a60e1d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.537529] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] volume-7f04fa3d-ec3b-4151-a46b-f232180344b9/volume-7f04fa3d-ec3b-4151-a46b-f232180344b9.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.538550] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333275, 'name': ReconfigVM_Task, 'duration_secs': 0.462104} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.542896] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c6de2ae-6a37-4065-9eef-e47f992c3634 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.557929] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 7504dfa6-bedc-4701-b4fc-60e19e742276/7504dfa6-bedc-4701-b4fc-60e19e742276.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 892.562809] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c29199fa-f6a9-4f4d-95aa-495f8e7eb919 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.569163] env[61906]: INFO nova.compute.manager [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Took 27.31 seconds to build instance. [ 892.574662] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333276, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.552} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.577451] env[61906]: INFO nova.virt.vmwareapi.ds_util [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 75b9738e-e9b6-435a-88bb-851982828d36/a9502563-1f96-46a5-b87e-12c0e9375e6e-rescue.vmdk. [ 892.578174] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 892.578174] env[61906]: value = "task-1333277" [ 892.578174] env[61906]: _type = "Task" [ 892.578174] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.578486] env[61906]: DEBUG oslo_vmware.api [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Waiting for the task: (returnval){ [ 892.578486] env[61906]: value = "task-1333278" [ 892.578486] env[61906]: _type = "Task" [ 892.578486] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.579285] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10d326f8-8b50-44ba-b028-959a321c3fbe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.595523] env[61906]: DEBUG oslo_vmware.api [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Task: {'id': task-1333278, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.619313] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 75b9738e-e9b6-435a-88bb-851982828d36/a9502563-1f96-46a5-b87e-12c0e9375e6e-rescue.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.619661] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333277, 'name': Rename_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.619923] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f4108431-30c4-420e-b057-3b343cfc460b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.638815] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Waiting for the task: (returnval){ [ 892.638815] env[61906]: value = "task-1333279" [ 892.638815] env[61906]: _type = "Task" [ 892.638815] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.648373] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333279, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.661699] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bfba3d6f-a2ca-45cf-ad53-3eee1f5394b4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.670975] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50d57ceb-9008-4601-b1df-1fc471041523 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.705134] env[61906]: DEBUG nova.compute.manager [req-c6ea175b-5c7d-44eb-b637-cf9642ee691b req-0f4a1f55-4a5c-4792-81d5-9c852962ffd8 service nova] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Detach interface failed, port_id=61a42623-e2e5-4245-9560-be57af9a8397, reason: Instance 9f1253e6-4727-4dca-bad2-04f0c5424cdb could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 892.766768] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "cb34a0ce-d274-4167-8367-e19809cc9e4d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.767070] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "cb34a0ce-d274-4167-8367-e19809cc9e4d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.767328] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "cb34a0ce-d274-4167-8367-e19809cc9e4d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.767523] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "cb34a0ce-d274-4167-8367-e19809cc9e4d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.767750] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "cb34a0ce-d274-4167-8367-e19809cc9e4d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.770411] env[61906]: INFO nova.compute.manager [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Terminating instance [ 892.772681] env[61906]: DEBUG nova.compute.manager [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 892.772896] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 892.773854] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9c9d48b-0349-4783-a019-937c9179ead5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.781886] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 892.781886] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-52bac208-9b7e-4c10-b678-44b46a133b8b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.788506] env[61906]: DEBUG oslo_vmware.api [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 892.788506] env[61906]: value = "task-1333280" [ 892.788506] env[61906]: _type = "Task" [ 892.788506] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.799427] env[61906]: DEBUG oslo_vmware.api [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333280, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.956922] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 893.071451] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4bb202e-b109-4952-87c0-c89e83220329 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "cc8870d0-eafb-4f73-aa79-b98f51370237" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.654s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.097929] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333277, 'name': Rename_Task, 'duration_secs': 0.31043} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.101974] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 893.102398] env[61906]: DEBUG oslo_vmware.api [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Task: {'id': task-1333278, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.104934] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-479ad6b2-545f-4344-a19b-4f35d43210f7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.111190] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 893.111190] env[61906]: value = "task-1333281" [ 893.111190] env[61906]: _type = "Task" [ 893.111190] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.119183] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333281, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.151699] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333279, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.194984] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acbe1575-fca2-4d09-b21f-05160c2dff65 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.204408] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-989b8881-8e40-47f2-80e3-c9075890c968 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.239059] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76877865-eb25-49d0-a947-2ff9d3f41780 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.247272] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e129755-c33e-44e2-a133-f37a003d2147 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.263796] env[61906]: DEBUG nova.compute.provider_tree [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 893.297876] env[61906]: DEBUG oslo_vmware.api [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333280, 'name': PowerOffVM_Task, 'duration_secs': 0.341874} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.298176] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 893.298358] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 893.298600] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e408903f-c026-44b1-bdee-456ee7cd4806 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.464315] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 893.464556] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Starting heal instance info cache {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 893.464727] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Rebuilding the list of instances to heal {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 893.471437] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 893.471702] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 893.471914] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Deleting the datastore file [datastore1] cb34a0ce-d274-4167-8367-e19809cc9e4d {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 893.472314] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4a635be3-7366-4ae7-b625-f3a545b67dae {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.480719] env[61906]: DEBUG oslo_vmware.api [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 893.480719] env[61906]: value = "task-1333283" [ 893.480719] env[61906]: _type = "Task" [ 893.480719] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.491030] env[61906]: DEBUG oslo_vmware.api [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333283, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.545349] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c5c1bb74-163e-4bde-ba10-ba195ccd6708 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "cc8870d0-eafb-4f73-aa79-b98f51370237" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.545834] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c5c1bb74-163e-4bde-ba10-ba195ccd6708 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "cc8870d0-eafb-4f73-aa79-b98f51370237" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.546133] env[61906]: DEBUG nova.compute.manager [None req-c5c1bb74-163e-4bde-ba10-ba195ccd6708 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 893.547112] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c40fe53-d97d-43d3-94dd-b06245099208 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.555952] env[61906]: DEBUG nova.compute.manager [None req-c5c1bb74-163e-4bde-ba10-ba195ccd6708 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61906) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 893.556814] env[61906]: DEBUG nova.objects.instance [None req-c5c1bb74-163e-4bde-ba10-ba195ccd6708 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lazy-loading 'flavor' on Instance uuid cc8870d0-eafb-4f73-aa79-b98f51370237 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.594111] env[61906]: DEBUG oslo_vmware.api [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Task: {'id': task-1333278, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.620119] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333281, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.651133] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333279, 'name': ReconfigVM_Task, 'duration_secs': 0.668235} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.651545] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 75b9738e-e9b6-435a-88bb-851982828d36/a9502563-1f96-46a5-b87e-12c0e9375e6e-rescue.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 893.652519] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-408c9126-7db8-4c61-8e51-b91dada95245 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.680780] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-253caba2-4e30-496f-9818-a2d60a6f454e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.696479] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Waiting for the task: (returnval){ [ 893.696479] env[61906]: value = "task-1333284" [ 893.696479] env[61906]: _type = "Task" [ 893.696479] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.709501] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333284, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.767432] env[61906]: DEBUG nova.scheduler.client.report [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 893.970512] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Skipping network cache update for instance because it is being deleted. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 893.970774] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 893.970987] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 893.971197] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 893.989912] env[61906]: DEBUG oslo_vmware.api [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333283, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.201621} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.990756] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "refresh_cache-9f1253e6-4727-4dca-bad2-04f0c5424cdb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.990891] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquired lock "refresh_cache-9f1253e6-4727-4dca-bad2-04f0c5424cdb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.991048] env[61906]: DEBUG nova.network.neutron [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Forcefully refreshing network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 893.991200] env[61906]: DEBUG nova.objects.instance [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lazy-loading 'info_cache' on Instance uuid 9f1253e6-4727-4dca-bad2-04f0c5424cdb {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.992501] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 893.992693] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 893.992878] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 893.993078] env[61906]: INFO nova.compute.manager [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Took 1.22 seconds to destroy the instance on the hypervisor. [ 893.993359] env[61906]: DEBUG oslo.service.loopingcall [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 893.993734] env[61906]: DEBUG nova.compute.manager [-] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 893.993829] env[61906]: DEBUG nova.network.neutron [-] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 894.064172] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5c1bb74-163e-4bde-ba10-ba195ccd6708 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 894.064752] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-782426f7-08f6-4c01-a5db-eefb6b801162 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.087763] env[61906]: DEBUG oslo_vmware.api [None req-c5c1bb74-163e-4bde-ba10-ba195ccd6708 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 894.087763] env[61906]: value = "task-1333285" [ 894.087763] env[61906]: _type = "Task" [ 894.087763] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.096860] env[61906]: DEBUG oslo_vmware.api [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Task: {'id': task-1333278, 'name': ReconfigVM_Task, 'duration_secs': 1.42021} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.099830] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Reconfigured VM instance instance-00000047 to attach disk [datastore2] volume-7f04fa3d-ec3b-4151-a46b-f232180344b9/volume-7f04fa3d-ec3b-4151-a46b-f232180344b9.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.104547] env[61906]: DEBUG oslo_vmware.api [None req-c5c1bb74-163e-4bde-ba10-ba195ccd6708 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333285, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.104797] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fef7660b-1967-4906-8b90-a8ba6e6c2dda {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.123154] env[61906]: DEBUG oslo_vmware.api [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333281, 'name': PowerOnVM_Task, 'duration_secs': 0.711663} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.124465] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 894.124691] env[61906]: INFO nova.compute.manager [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Took 8.82 seconds to spawn the instance on the hypervisor. [ 894.124874] env[61906]: DEBUG nova.compute.manager [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 894.125222] env[61906]: DEBUG oslo_vmware.api [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Waiting for the task: (returnval){ [ 894.125222] env[61906]: value = "task-1333286" [ 894.125222] env[61906]: _type = "Task" [ 894.125222] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.125917] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edbf34db-f9df-4543-bd29-90a9b88c1047 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.141714] env[61906]: DEBUG oslo_vmware.api [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Task: {'id': task-1333286, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.207639] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333284, 'name': ReconfigVM_Task, 'duration_secs': 0.194727} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.207928] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 894.208791] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8dcfba24-fa19-4bc7-a896-b07ece0df630 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.215939] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Waiting for the task: (returnval){ [ 894.215939] env[61906]: value = "task-1333287" [ 894.215939] env[61906]: _type = "Task" [ 894.215939] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.224442] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333287, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.276261] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.315s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.276261] env[61906]: DEBUG nova.compute.manager [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 894.278367] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.935s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.280778] env[61906]: INFO nova.compute.claims [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 894.316248] env[61906]: DEBUG nova.compute.manager [req-213fe092-f1b5-479e-80c5-013edc1b92e1 req-4c6736f1-9f33-406b-b397-bcb6db11bb64 service nova] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Received event network-vif-deleted-f23a470d-0c5d-4aca-95ea-22c1ad18ca49 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 894.316248] env[61906]: INFO nova.compute.manager [req-213fe092-f1b5-479e-80c5-013edc1b92e1 req-4c6736f1-9f33-406b-b397-bcb6db11bb64 service nova] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Neutron deleted interface f23a470d-0c5d-4aca-95ea-22c1ad18ca49; detaching it from the instance and deleting it from the info cache [ 894.316248] env[61906]: DEBUG nova.network.neutron [req-213fe092-f1b5-479e-80c5-013edc1b92e1 req-4c6736f1-9f33-406b-b397-bcb6db11bb64 service nova] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.545403] env[61906]: DEBUG oslo_vmware.rw_handles [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5264888f-acba-b2b2-38e4-d1a6b31affde/disk-0.vmdk. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 894.546448] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2898e32-cef3-4d50-8b85-499c7ea0d6fc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.553633] env[61906]: DEBUG oslo_vmware.rw_handles [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5264888f-acba-b2b2-38e4-d1a6b31affde/disk-0.vmdk is in state: ready. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 894.553835] env[61906]: ERROR oslo_vmware.rw_handles [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5264888f-acba-b2b2-38e4-d1a6b31affde/disk-0.vmdk due to incomplete transfer. [ 894.554080] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-d0812e89-1f97-436c-915e-e3d24c7a0198 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.561807] env[61906]: DEBUG oslo_vmware.rw_handles [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5264888f-acba-b2b2-38e4-d1a6b31affde/disk-0.vmdk. {{(pid=61906) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 894.562038] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Uploaded image 9ea4f3cc-c567-4887-b2ca-8874462a07fc to the Glance image server {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 894.564330] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Destroying the VM {{(pid=61906) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 894.564597] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-4ea53eef-5023-4d24-b2e5-34aa1acb9ed3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.569967] env[61906]: DEBUG oslo_vmware.api [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Waiting for the task: (returnval){ [ 894.569967] env[61906]: value = "task-1333288" [ 894.569967] env[61906]: _type = "Task" [ 894.569967] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.580149] env[61906]: DEBUG oslo_vmware.api [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333288, 'name': Destroy_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.597253] env[61906]: DEBUG oslo_vmware.api [None req-c5c1bb74-163e-4bde-ba10-ba195ccd6708 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333285, 'name': PowerOffVM_Task, 'duration_secs': 0.228698} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.597518] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5c1bb74-163e-4bde-ba10-ba195ccd6708 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 894.597694] env[61906]: DEBUG nova.compute.manager [None req-c5c1bb74-163e-4bde-ba10-ba195ccd6708 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 894.598468] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4b7385f-343f-4698-9235-d8c9da343ecf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.637127] env[61906]: DEBUG oslo_vmware.api [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Task: {'id': task-1333286, 'name': ReconfigVM_Task, 'duration_secs': 0.18954} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.637786] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284737', 'volume_id': '7f04fa3d-ec3b-4151-a46b-f232180344b9', 'name': 'volume-7f04fa3d-ec3b-4151-a46b-f232180344b9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'da493512-d996-4de7-9f47-cadcbc4fbcb9', 'attached_at': '', 'detached_at': '', 'volume_id': '7f04fa3d-ec3b-4151-a46b-f232180344b9', 'serial': '7f04fa3d-ec3b-4151-a46b-f232180344b9'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 894.638840] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8e12a858-f93c-4813-ad94-5f3a47ae55b0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.647857] env[61906]: DEBUG oslo_vmware.api [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Waiting for the task: (returnval){ [ 894.647857] env[61906]: value = "task-1333289" [ 894.647857] env[61906]: _type = "Task" [ 894.647857] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.652558] env[61906]: INFO nova.compute.manager [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Took 27.85 seconds to build instance. [ 894.659423] env[61906]: DEBUG oslo_vmware.api [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Task: {'id': task-1333289, 'name': Rename_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.725946] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333287, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.789252] env[61906]: DEBUG nova.compute.utils [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 894.792549] env[61906]: DEBUG nova.compute.manager [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 894.792819] env[61906]: DEBUG nova.network.neutron [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 894.795012] env[61906]: DEBUG nova.network.neutron [-] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.818817] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-87b317a3-ccae-40ca-8963-a2e3abc58ae5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.828929] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76168bd7-0232-4d67-b48b-d5ef17793334 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.866401] env[61906]: DEBUG nova.compute.manager [req-213fe092-f1b5-479e-80c5-013edc1b92e1 req-4c6736f1-9f33-406b-b397-bcb6db11bb64 service nova] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Detach interface failed, port_id=f23a470d-0c5d-4aca-95ea-22c1ad18ca49, reason: Instance cb34a0ce-d274-4167-8367-e19809cc9e4d could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 894.899327] env[61906]: DEBUG nova.policy [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fd25149493b948f2826e6688c9958e6a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4fbdcd2dc0d04e83b9f82a19bdfd0023', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 895.023106] env[61906]: DEBUG nova.network.neutron [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 895.079466] env[61906]: DEBUG oslo_vmware.api [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333288, 'name': Destroy_Task} progress is 33%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.110505] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c5c1bb74-163e-4bde-ba10-ba195ccd6708 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "cc8870d0-eafb-4f73-aa79-b98f51370237" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.565s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.155238] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ee13f901-3ca2-441b-bcb0-b6b6d29ffeb7 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "7504dfa6-bedc-4701-b4fc-60e19e742276" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.421s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.161316] env[61906]: DEBUG oslo_vmware.api [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Task: {'id': task-1333289, 'name': Rename_Task, 'duration_secs': 0.34741} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.161810] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 895.162113] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8ec6cf67-e436-4bb9-a9da-131c8405ebb5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.169092] env[61906]: DEBUG oslo_vmware.api [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Waiting for the task: (returnval){ [ 895.169092] env[61906]: value = "task-1333290" [ 895.169092] env[61906]: _type = "Task" [ 895.169092] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.180019] env[61906]: DEBUG oslo_vmware.api [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Task: {'id': task-1333290, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.226375] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333287, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.296096] env[61906]: DEBUG nova.compute.manager [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 895.304644] env[61906]: INFO nova.compute.manager [-] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Took 1.31 seconds to deallocate network for instance. [ 895.419073] env[61906]: DEBUG nova.network.neutron [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Successfully created port: ba5a12df-7917-4b2d-ac9e-7ef1380dabf7 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 895.547059] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69eb8ce3-b696-4aa2-85aa-de9c7a30fd7b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.556803] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fae79df-136e-4b26-9057-a211e40e137d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.595294] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fffc41f8-24c1-4177-8480-54f47377d219 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.603260] env[61906]: DEBUG oslo_vmware.api [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333288, 'name': Destroy_Task, 'duration_secs': 0.681264} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.608938] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Destroyed the VM [ 895.609320] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Deleting Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 895.609664] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-fde4f69f-5edd-4412-a6f9-494202a19d7d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.612864] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4267c4da-aa2a-4771-9226-53d650830d57 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.628671] env[61906]: DEBUG nova.compute.provider_tree [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.634474] env[61906]: DEBUG oslo_vmware.api [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Waiting for the task: (returnval){ [ 895.634474] env[61906]: value = "task-1333291" [ 895.634474] env[61906]: _type = "Task" [ 895.634474] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.641932] env[61906]: DEBUG oslo_vmware.api [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333291, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.681855] env[61906]: DEBUG oslo_vmware.api [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Task: {'id': task-1333290, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.706494] env[61906]: DEBUG nova.network.neutron [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.727220] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333287, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.812402] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.133562] env[61906]: DEBUG nova.scheduler.client.report [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 896.145619] env[61906]: DEBUG oslo_vmware.api [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333291, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.178797] env[61906]: DEBUG oslo_vmware.api [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Task: {'id': task-1333290, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.208948] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Releasing lock "refresh_cache-9f1253e6-4727-4dca-bad2-04f0c5424cdb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.209195] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Updated the network info_cache for instance {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 896.209410] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 896.209567] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 896.209719] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 896.209861] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 896.226019] env[61906]: DEBUG oslo_vmware.api [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333287, 'name': PowerOnVM_Task, 'duration_secs': 1.559026} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.226360] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 896.229317] env[61906]: DEBUG nova.compute.manager [None req-750e0460-d288-4caa-beb2-26d773d4bad7 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.230087] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ea0e3c3-eab1-4377-8a93-6d8a5ab183ab {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.307116] env[61906]: DEBUG nova.compute.manager [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 896.332613] env[61906]: DEBUG nova.virt.hardware [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 896.332895] env[61906]: DEBUG nova.virt.hardware [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 896.333158] env[61906]: DEBUG nova.virt.hardware [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 896.333394] env[61906]: DEBUG nova.virt.hardware [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 896.333563] env[61906]: DEBUG nova.virt.hardware [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 896.333729] env[61906]: DEBUG nova.virt.hardware [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 896.333960] env[61906]: DEBUG nova.virt.hardware [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 896.334203] env[61906]: DEBUG nova.virt.hardware [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 896.334385] env[61906]: DEBUG nova.virt.hardware [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 896.335028] env[61906]: DEBUG nova.virt.hardware [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 896.335028] env[61906]: DEBUG nova.virt.hardware [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 896.335810] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-461ac078-8b53-43dc-abc1-8d2fdab13580 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.348884] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49447fd0-4f7b-4ea9-9516-9f4eea79a89d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.355110] env[61906]: DEBUG nova.compute.manager [req-bb94634f-ef9d-45d0-883b-c7bf9bda4116 req-a145c966-7081-4acd-a30a-5f9640c19e85 service nova] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Received event network-changed-175a521b-1c70-41f6-963c-79c978013fa1 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 896.355316] env[61906]: DEBUG nova.compute.manager [req-bb94634f-ef9d-45d0-883b-c7bf9bda4116 req-a145c966-7081-4acd-a30a-5f9640c19e85 service nova] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Refreshing instance network info cache due to event network-changed-175a521b-1c70-41f6-963c-79c978013fa1. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 896.355539] env[61906]: DEBUG oslo_concurrency.lockutils [req-bb94634f-ef9d-45d0-883b-c7bf9bda4116 req-a145c966-7081-4acd-a30a-5f9640c19e85 service nova] Acquiring lock "refresh_cache-7504dfa6-bedc-4701-b4fc-60e19e742276" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.355686] env[61906]: DEBUG oslo_concurrency.lockutils [req-bb94634f-ef9d-45d0-883b-c7bf9bda4116 req-a145c966-7081-4acd-a30a-5f9640c19e85 service nova] Acquired lock "refresh_cache-7504dfa6-bedc-4701-b4fc-60e19e742276" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.355852] env[61906]: DEBUG nova.network.neutron [req-bb94634f-ef9d-45d0-883b-c7bf9bda4116 req-a145c966-7081-4acd-a30a-5f9640c19e85 service nova] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Refreshing network info cache for port 175a521b-1c70-41f6-963c-79c978013fa1 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 896.643081] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.364s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.643253] env[61906]: DEBUG nova.compute.manager [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 896.646395] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.712s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.648457] env[61906]: INFO nova.compute.claims [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 896.658742] env[61906]: DEBUG oslo_vmware.api [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333291, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.681046] env[61906]: DEBUG oslo_vmware.api [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Task: {'id': task-1333290, 'name': PowerOnVM_Task} progress is 91%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.908733] env[61906]: DEBUG nova.compute.manager [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.909653] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5873812-c579-47a4-b892-a02d2e21774a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.026640] env[61906]: DEBUG nova.compute.manager [req-b4d01932-a754-44b6-b158-00fe828b6391 req-4cb0aaf4-6309-4093-b704-f2d75ea34200 service nova] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Received event network-vif-plugged-ba5a12df-7917-4b2d-ac9e-7ef1380dabf7 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 897.026934] env[61906]: DEBUG oslo_concurrency.lockutils [req-b4d01932-a754-44b6-b158-00fe828b6391 req-4cb0aaf4-6309-4093-b704-f2d75ea34200 service nova] Acquiring lock "b1159533-c970-49d6-af42-b954b20d92fb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.027137] env[61906]: DEBUG oslo_concurrency.lockutils [req-b4d01932-a754-44b6-b158-00fe828b6391 req-4cb0aaf4-6309-4093-b704-f2d75ea34200 service nova] Lock "b1159533-c970-49d6-af42-b954b20d92fb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.027299] env[61906]: DEBUG oslo_concurrency.lockutils [req-b4d01932-a754-44b6-b158-00fe828b6391 req-4cb0aaf4-6309-4093-b704-f2d75ea34200 service nova] Lock "b1159533-c970-49d6-af42-b954b20d92fb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.027452] env[61906]: DEBUG nova.compute.manager [req-b4d01932-a754-44b6-b158-00fe828b6391 req-4cb0aaf4-6309-4093-b704-f2d75ea34200 service nova] [instance: b1159533-c970-49d6-af42-b954b20d92fb] No waiting events found dispatching network-vif-plugged-ba5a12df-7917-4b2d-ac9e-7ef1380dabf7 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 897.027615] env[61906]: WARNING nova.compute.manager [req-b4d01932-a754-44b6-b158-00fe828b6391 req-4cb0aaf4-6309-4093-b704-f2d75ea34200 service nova] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Received unexpected event network-vif-plugged-ba5a12df-7917-4b2d-ac9e-7ef1380dabf7 for instance with vm_state building and task_state spawning. [ 897.104707] env[61906]: DEBUG nova.network.neutron [req-bb94634f-ef9d-45d0-883b-c7bf9bda4116 req-a145c966-7081-4acd-a30a-5f9640c19e85 service nova] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Updated VIF entry in instance network info cache for port 175a521b-1c70-41f6-963c-79c978013fa1. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 897.104707] env[61906]: DEBUG nova.network.neutron [req-bb94634f-ef9d-45d0-883b-c7bf9bda4116 req-a145c966-7081-4acd-a30a-5f9640c19e85 service nova] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Updating instance_info_cache with network_info: [{"id": "175a521b-1c70-41f6-963c-79c978013fa1", "address": "fa:16:3e:dd:31:4c", "network": {"id": "c3bcef0f-44a7-44f7-8130-df77c0d3ac2d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-330384233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.155", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "129cdb06b47346b6895df45e73f7b2dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35e463c7-7d78-4d66-8efd-6127b1f3ee17", "external-id": "nsx-vlan-transportzone-175", "segmentation_id": 175, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap175a521b-1c", "ovs_interfaceid": "175a521b-1c70-41f6-963c-79c978013fa1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.151120] env[61906]: DEBUG oslo_vmware.api [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333291, 'name': RemoveSnapshot_Task, 'duration_secs': 1.082086} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.151800] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Deleted Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 897.152197] env[61906]: INFO nova.compute.manager [None req-75eba799-e3dc-487f-8fd7-6570df0ea0f1 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Took 17.00 seconds to snapshot the instance on the hypervisor. [ 897.157094] env[61906]: DEBUG nova.compute.utils [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 897.164108] env[61906]: DEBUG nova.compute.manager [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 897.164108] env[61906]: DEBUG nova.network.neutron [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 897.181743] env[61906]: DEBUG oslo_vmware.api [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Task: {'id': task-1333290, 'name': PowerOnVM_Task, 'duration_secs': 1.539918} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.183409] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 897.183409] env[61906]: INFO nova.compute.manager [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Took 7.71 seconds to spawn the instance on the hypervisor. [ 897.183409] env[61906]: DEBUG nova.compute.manager [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 897.183780] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d05082e-05e4-4b77-9d34-1155d0f01ba7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.206547] env[61906]: DEBUG nova.policy [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e89d45df95134c709b569d06744dea5e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1b33dea0794a48f78b9f519cb269a8c2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 897.432890] env[61906]: INFO nova.compute.manager [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] instance snapshotting [ 897.433935] env[61906]: WARNING nova.compute.manager [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] trying to snapshot a non-running instance: (state: 4 expected: 1) [ 897.437014] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1977bdc3-0a75-4cc7-9b50-ec7f44e48c2b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.459118] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f19b4ea4-e5cd-4207-a6fc-37d178c56f7f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.472564] env[61906]: DEBUG nova.network.neutron [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Successfully updated port: ba5a12df-7917-4b2d-ac9e-7ef1380dabf7 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 897.606577] env[61906]: DEBUG oslo_concurrency.lockutils [req-bb94634f-ef9d-45d0-883b-c7bf9bda4116 req-a145c966-7081-4acd-a30a-5f9640c19e85 service nova] Releasing lock "refresh_cache-7504dfa6-bedc-4701-b4fc-60e19e742276" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.667394] env[61906]: DEBUG nova.compute.manager [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 897.704244] env[61906]: INFO nova.compute.manager [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Took 30.61 seconds to build instance. [ 897.751295] env[61906]: DEBUG nova.network.neutron [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Successfully created port: 1af3a42f-34db-4c40-8efc-330a284d1f2c {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 897.913567] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2128f863-9227-4b81-b998-8434ea8af7c9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.921664] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12c8aa3c-1332-453c-b24b-363ad99385cd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.953517] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-477f1b4b-8f60-48c4-81be-9c41c0f0bb53 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.961519] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9119654d-f225-480c-9aa9-bbe14c887e02 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.978126] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Creating Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 897.978692] env[61906]: DEBUG nova.compute.provider_tree [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 897.980075] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-a3918627-33d8-4b72-be00-33571b338af4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.982582] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "refresh_cache-b1159533-c970-49d6-af42-b954b20d92fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.982582] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquired lock "refresh_cache-b1159533-c970-49d6-af42-b954b20d92fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.985016] env[61906]: DEBUG nova.network.neutron [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 897.995343] env[61906]: DEBUG oslo_vmware.api [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 897.995343] env[61906]: value = "task-1333292" [ 897.995343] env[61906]: _type = "Task" [ 897.995343] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.005679] env[61906]: DEBUG oslo_vmware.api [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333292, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.206184] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d4b41d2f-41da-4a0b-a048-d2b5711dfaf0 tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Lock "da493512-d996-4de7-9f47-cadcbc4fbcb9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.837s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.429446] env[61906]: DEBUG nova.compute.manager [req-6f60082f-2359-44a1-a1ed-ef04f089952f req-f9e9a2d8-74df-48fe-a1d8-43a0fadc674f service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Received event network-changed-8bcec3ac-5104-4e72-af50-08fae9dc156c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 898.429663] env[61906]: DEBUG nova.compute.manager [req-6f60082f-2359-44a1-a1ed-ef04f089952f req-f9e9a2d8-74df-48fe-a1d8-43a0fadc674f service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Refreshing instance network info cache due to event network-changed-8bcec3ac-5104-4e72-af50-08fae9dc156c. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 898.429885] env[61906]: DEBUG oslo_concurrency.lockutils [req-6f60082f-2359-44a1-a1ed-ef04f089952f req-f9e9a2d8-74df-48fe-a1d8-43a0fadc674f service nova] Acquiring lock "refresh_cache-75b9738e-e9b6-435a-88bb-851982828d36" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.430075] env[61906]: DEBUG oslo_concurrency.lockutils [req-6f60082f-2359-44a1-a1ed-ef04f089952f req-f9e9a2d8-74df-48fe-a1d8-43a0fadc674f service nova] Acquired lock "refresh_cache-75b9738e-e9b6-435a-88bb-851982828d36" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.430274] env[61906]: DEBUG nova.network.neutron [req-6f60082f-2359-44a1-a1ed-ef04f089952f req-f9e9a2d8-74df-48fe-a1d8-43a0fadc674f service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Refreshing network info cache for port 8bcec3ac-5104-4e72-af50-08fae9dc156c {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 898.489134] env[61906]: DEBUG nova.scheduler.client.report [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 898.507611] env[61906]: DEBUG oslo_vmware.api [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333292, 'name': CreateSnapshot_Task, 'duration_secs': 0.410308} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.508131] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Created Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 898.508888] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd041644-2afc-42e4-b1f4-e782ac12ebf9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.569071] env[61906]: DEBUG nova.network.neutron [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 898.678653] env[61906]: DEBUG nova.compute.manager [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 898.703470] env[61906]: DEBUG nova.virt.hardware [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 898.703724] env[61906]: DEBUG nova.virt.hardware [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 898.703898] env[61906]: DEBUG nova.virt.hardware [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 898.704106] env[61906]: DEBUG nova.virt.hardware [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 898.704263] env[61906]: DEBUG nova.virt.hardware [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 898.704425] env[61906]: DEBUG nova.virt.hardware [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 898.704653] env[61906]: DEBUG nova.virt.hardware [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 898.704824] env[61906]: DEBUG nova.virt.hardware [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 898.704995] env[61906]: DEBUG nova.virt.hardware [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 898.705283] env[61906]: DEBUG nova.virt.hardware [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 898.705468] env[61906]: DEBUG nova.virt.hardware [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 898.706341] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64673e30-868b-4e10-9bc9-22b237caae12 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.716638] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c16a47d1-dbf2-444e-bfb7-7a5cea5d9062 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.797131] env[61906]: DEBUG nova.compute.manager [req-93bd2ae9-2dc6-44bc-9c57-efb99b987104 req-df025e8d-b11f-4410-84b1-e390f92cc71c service nova] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Received event network-changed-ba5a12df-7917-4b2d-ac9e-7ef1380dabf7 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 898.797476] env[61906]: DEBUG nova.compute.manager [req-93bd2ae9-2dc6-44bc-9c57-efb99b987104 req-df025e8d-b11f-4410-84b1-e390f92cc71c service nova] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Refreshing instance network info cache due to event network-changed-ba5a12df-7917-4b2d-ac9e-7ef1380dabf7. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 898.797597] env[61906]: DEBUG oslo_concurrency.lockutils [req-93bd2ae9-2dc6-44bc-9c57-efb99b987104 req-df025e8d-b11f-4410-84b1-e390f92cc71c service nova] Acquiring lock "refresh_cache-b1159533-c970-49d6-af42-b954b20d92fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.922458] env[61906]: DEBUG nova.network.neutron [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Updating instance_info_cache with network_info: [{"id": "ba5a12df-7917-4b2d-ac9e-7ef1380dabf7", "address": "fa:16:3e:d3:84:8a", "network": {"id": "1c55d0cc-4c30-4b89-8736-3e96c5c9c507", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-245940583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fbdcd2dc0d04e83b9f82a19bdfd0023", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bcd9d2d-25c8-41ad-9a4a-93b9029ba993", "external-id": "nsx-vlan-transportzone-734", "segmentation_id": 734, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba5a12df-79", "ovs_interfaceid": "ba5a12df-7917-4b2d-ac9e-7ef1380dabf7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.996739] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.350s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.997344] env[61906]: DEBUG nova.compute.manager [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 899.000512] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.527s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.000721] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.002843] env[61906]: DEBUG oslo_concurrency.lockutils [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 11.402s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.003020] env[61906]: DEBUG nova.objects.instance [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61906) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 899.027860] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Creating linked-clone VM from snapshot {{(pid=61906) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 899.027860] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8cf90da9-6c74-4446-9b78-402afe0477b0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.037939] env[61906]: DEBUG oslo_vmware.api [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 899.037939] env[61906]: value = "task-1333293" [ 899.037939] env[61906]: _type = "Task" [ 899.037939] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.045058] env[61906]: INFO nova.scheduler.client.report [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Deleted allocations for instance 80a6a0b8-17c3-46e8-8861-6d1759a19111 [ 899.055238] env[61906]: DEBUG oslo_vmware.api [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333293, 'name': CloneVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.104886] env[61906]: DEBUG nova.compute.manager [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 899.106709] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35c8b69b-f85d-449d-85c4-1c21b8036f39 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.354263] env[61906]: DEBUG nova.network.neutron [req-6f60082f-2359-44a1-a1ed-ef04f089952f req-f9e9a2d8-74df-48fe-a1d8-43a0fadc674f service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Updated VIF entry in instance network info cache for port 8bcec3ac-5104-4e72-af50-08fae9dc156c. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 899.354741] env[61906]: DEBUG nova.network.neutron [req-6f60082f-2359-44a1-a1ed-ef04f089952f req-f9e9a2d8-74df-48fe-a1d8-43a0fadc674f service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Updating instance_info_cache with network_info: [{"id": "8bcec3ac-5104-4e72-af50-08fae9dc156c", "address": "fa:16:3e:36:f8:fd", "network": {"id": "eaf64f00-dd3b-41c3-b679-3ee2f53735a0", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-255429219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "5375ff2ae28b4427a3c4541ee3d2729a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20fbc86c-0935-406c-9659-3ffc33fc0d08", "external-id": "nsx-vlan-transportzone-191", "segmentation_id": 191, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bcec3ac-51", "ovs_interfaceid": "8bcec3ac-5104-4e72-af50-08fae9dc156c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.425559] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Releasing lock "refresh_cache-b1159533-c970-49d6-af42-b954b20d92fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.426060] env[61906]: DEBUG nova.compute.manager [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Instance network_info: |[{"id": "ba5a12df-7917-4b2d-ac9e-7ef1380dabf7", "address": "fa:16:3e:d3:84:8a", "network": {"id": "1c55d0cc-4c30-4b89-8736-3e96c5c9c507", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-245940583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fbdcd2dc0d04e83b9f82a19bdfd0023", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bcd9d2d-25c8-41ad-9a4a-93b9029ba993", "external-id": "nsx-vlan-transportzone-734", "segmentation_id": 734, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba5a12df-79", "ovs_interfaceid": "ba5a12df-7917-4b2d-ac9e-7ef1380dabf7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 899.426548] env[61906]: DEBUG oslo_concurrency.lockutils [req-93bd2ae9-2dc6-44bc-9c57-efb99b987104 req-df025e8d-b11f-4410-84b1-e390f92cc71c service nova] Acquired lock "refresh_cache-b1159533-c970-49d6-af42-b954b20d92fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.426794] env[61906]: DEBUG nova.network.neutron [req-93bd2ae9-2dc6-44bc-9c57-efb99b987104 req-df025e8d-b11f-4410-84b1-e390f92cc71c service nova] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Refreshing network info cache for port ba5a12df-7917-4b2d-ac9e-7ef1380dabf7 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 899.431290] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d3:84:8a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7bcd9d2d-25c8-41ad-9a4a-93b9029ba993', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ba5a12df-7917-4b2d-ac9e-7ef1380dabf7', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 899.436711] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Creating folder: Project (4fbdcd2dc0d04e83b9f82a19bdfd0023). Parent ref: group-v284713. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 899.437734] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-98f747c7-97cd-47ce-af0a-96200581153d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.452669] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Created folder: Project (4fbdcd2dc0d04e83b9f82a19bdfd0023) in parent group-v284713. [ 899.452898] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Creating folder: Instances. Parent ref: group-v284775. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 899.453185] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5977cd55-7fd4-44c7-83eb-bcdbac79253b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.463967] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Created folder: Instances in parent group-v284775. [ 899.464270] env[61906]: DEBUG oslo.service.loopingcall [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 899.464506] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 899.464708] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ab0ac133-c5c9-4a55-87b7-f96b35fe8ea6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.488019] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 899.488019] env[61906]: value = "task-1333296" [ 899.488019] env[61906]: _type = "Task" [ 899.488019] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.497490] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333296, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.509253] env[61906]: DEBUG nova.compute.utils [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 899.513651] env[61906]: DEBUG nova.compute.manager [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 899.513777] env[61906]: DEBUG nova.network.neutron [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 899.550323] env[61906]: DEBUG oslo_vmware.api [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333293, 'name': CloneVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.557033] env[61906]: DEBUG nova.network.neutron [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Successfully updated port: 1af3a42f-34db-4c40-8efc-330a284d1f2c {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 899.563149] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ccb9449b-854b-4161-ac23-23c13f6385e5 tempest-SecurityGroupsTestJSON-1707852659 tempest-SecurityGroupsTestJSON-1707852659-project-member] Lock "80a6a0b8-17c3-46e8-8861-6d1759a19111" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.988s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.587790] env[61906]: DEBUG nova.policy [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2eec83ef55eb44efaef3185e0773182b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '357662e4b2e7430b82a521ed0f4392ad', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 899.620720] env[61906]: INFO nova.compute.manager [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] instance snapshotting [ 899.625653] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99ba0398-ec2e-493f-9b25-7c0601b7d705 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.645920] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b3752d7-d77f-4ea7-9df0-da14153fe475 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.857411] env[61906]: DEBUG oslo_concurrency.lockutils [req-6f60082f-2359-44a1-a1ed-ef04f089952f req-f9e9a2d8-74df-48fe-a1d8-43a0fadc674f service nova] Releasing lock "refresh_cache-75b9738e-e9b6-435a-88bb-851982828d36" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.007347] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333296, 'name': CreateVM_Task, 'duration_secs': 0.360606} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.007595] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 900.008215] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.008465] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.008791] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 900.009067] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e7e12d5-1d55-4de2-b777-443d22d0dd72 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.015078] env[61906]: DEBUG oslo_vmware.api [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 900.015078] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52d2be07-5268-1b52-a846-a5c80b76cd76" [ 900.015078] env[61906]: _type = "Task" [ 900.015078] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.019268] env[61906]: DEBUG oslo_concurrency.lockutils [None req-601c2df8-062c-40f8-9c9b-c8d773745fa8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.020400] env[61906]: DEBUG nova.compute.manager [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 900.024386] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.280s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.024386] env[61906]: DEBUG nova.objects.instance [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Lazy-loading 'resources' on Instance uuid 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 900.043819] env[61906]: DEBUG oslo_vmware.api [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52d2be07-5268-1b52-a846-a5c80b76cd76, 'name': SearchDatastore_Task, 'duration_secs': 0.024627} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.049239] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.049707] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 900.051138] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.051138] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.051138] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 900.051606] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1f63ab4f-0d12-403c-8a11-657a24a90cc2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.063692] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "refresh_cache-1bc9f3d4-b539-4841-b5b6-eaefa0d06a72" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.064366] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquired lock "refresh_cache-1bc9f3d4-b539-4841-b5b6-eaefa0d06a72" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.065585] env[61906]: DEBUG nova.network.neutron [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 900.067418] env[61906]: DEBUG oslo_vmware.api [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333293, 'name': CloneVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.069272] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 900.069506] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 900.070708] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e868d4a-9ac2-4e98-a9b0-7625c3560d66 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.080636] env[61906]: DEBUG nova.network.neutron [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Successfully created port: 7e726302-5576-4a98-b2a5-23c4ed0be21e {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 900.085465] env[61906]: DEBUG oslo_vmware.api [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 900.085465] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5233a7fa-66ed-fefb-36d5-9d95f36742a9" [ 900.085465] env[61906]: _type = "Task" [ 900.085465] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.100049] env[61906]: DEBUG oslo_vmware.api [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5233a7fa-66ed-fefb-36d5-9d95f36742a9, 'name': SearchDatastore_Task, 'duration_secs': 0.01193} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.101067] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4706a509-6969-4dac-a852-3ac94012e347 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.110025] env[61906]: DEBUG oslo_vmware.api [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 900.110025] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]526ad90d-3c62-3e5d-1424-e9a0c201a753" [ 900.110025] env[61906]: _type = "Task" [ 900.110025] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.119352] env[61906]: DEBUG oslo_vmware.api [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]526ad90d-3c62-3e5d-1424-e9a0c201a753, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.157835] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Creating Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 900.158173] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f87e839a-44e5-423e-9207-7cee2be67d9d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.168737] env[61906]: DEBUG oslo_vmware.api [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Waiting for the task: (returnval){ [ 900.168737] env[61906]: value = "task-1333297" [ 900.168737] env[61906]: _type = "Task" [ 900.168737] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.177916] env[61906]: DEBUG oslo_vmware.api [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333297, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.274215] env[61906]: DEBUG nova.network.neutron [req-93bd2ae9-2dc6-44bc-9c57-efb99b987104 req-df025e8d-b11f-4410-84b1-e390f92cc71c service nova] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Updated VIF entry in instance network info cache for port ba5a12df-7917-4b2d-ac9e-7ef1380dabf7. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 900.274605] env[61906]: DEBUG nova.network.neutron [req-93bd2ae9-2dc6-44bc-9c57-efb99b987104 req-df025e8d-b11f-4410-84b1-e390f92cc71c service nova] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Updating instance_info_cache with network_info: [{"id": "ba5a12df-7917-4b2d-ac9e-7ef1380dabf7", "address": "fa:16:3e:d3:84:8a", "network": {"id": "1c55d0cc-4c30-4b89-8736-3e96c5c9c507", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-245940583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fbdcd2dc0d04e83b9f82a19bdfd0023", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bcd9d2d-25c8-41ad-9a4a-93b9029ba993", "external-id": "nsx-vlan-transportzone-734", "segmentation_id": 734, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba5a12df-79", "ovs_interfaceid": "ba5a12df-7917-4b2d-ac9e-7ef1380dabf7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.527447] env[61906]: DEBUG nova.compute.manager [req-f636d7bf-27a6-4f4b-997f-e3e0c158000f req-a5ab416c-3c5b-459a-947f-9c9c767ef98a service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Received event network-changed-8bcec3ac-5104-4e72-af50-08fae9dc156c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 900.527713] env[61906]: DEBUG nova.compute.manager [req-f636d7bf-27a6-4f4b-997f-e3e0c158000f req-a5ab416c-3c5b-459a-947f-9c9c767ef98a service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Refreshing instance network info cache due to event network-changed-8bcec3ac-5104-4e72-af50-08fae9dc156c. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 900.527971] env[61906]: DEBUG oslo_concurrency.lockutils [req-f636d7bf-27a6-4f4b-997f-e3e0c158000f req-a5ab416c-3c5b-459a-947f-9c9c767ef98a service nova] Acquiring lock "refresh_cache-75b9738e-e9b6-435a-88bb-851982828d36" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.528183] env[61906]: DEBUG oslo_concurrency.lockutils [req-f636d7bf-27a6-4f4b-997f-e3e0c158000f req-a5ab416c-3c5b-459a-947f-9c9c767ef98a service nova] Acquired lock "refresh_cache-75b9738e-e9b6-435a-88bb-851982828d36" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.528423] env[61906]: DEBUG nova.network.neutron [req-f636d7bf-27a6-4f4b-997f-e3e0c158000f req-a5ab416c-3c5b-459a-947f-9c9c767ef98a service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Refreshing network info cache for port 8bcec3ac-5104-4e72-af50-08fae9dc156c {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 900.562967] env[61906]: DEBUG oslo_vmware.api [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333293, 'name': CloneVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.623250] env[61906]: DEBUG oslo_vmware.api [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]526ad90d-3c62-3e5d-1424-e9a0c201a753, 'name': SearchDatastore_Task, 'duration_secs': 0.014539} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.624063] env[61906]: DEBUG nova.network.neutron [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 900.626861] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.626861] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] b1159533-c970-49d6-af42-b954b20d92fb/b1159533-c970-49d6-af42-b954b20d92fb.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 900.627101] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-14db1f9a-292a-4696-ad45-13807f173ccb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.636997] env[61906]: DEBUG oslo_vmware.api [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 900.636997] env[61906]: value = "task-1333298" [ 900.636997] env[61906]: _type = "Task" [ 900.636997] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.646529] env[61906]: DEBUG oslo_vmware.api [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333298, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.682611] env[61906]: DEBUG oslo_vmware.api [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333297, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.750889] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e2421c2-1ca8-4c76-86df-8802a7477d41 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.759191] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b0e9990-7cb0-4fa3-bd51-747258302df2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.794922] env[61906]: DEBUG oslo_concurrency.lockutils [req-93bd2ae9-2dc6-44bc-9c57-efb99b987104 req-df025e8d-b11f-4410-84b1-e390f92cc71c service nova] Releasing lock "refresh_cache-b1159533-c970-49d6-af42-b954b20d92fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.798734] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dff11e7-60d4-4a8c-9d94-2f724b00d79f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.807422] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61a50ec2-e0e9-49c1-9ea0-01803d1d4d8a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.822793] env[61906]: DEBUG nova.compute.provider_tree [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.846754] env[61906]: DEBUG nova.network.neutron [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Updating instance_info_cache with network_info: [{"id": "1af3a42f-34db-4c40-8efc-330a284d1f2c", "address": "fa:16:3e:39:df:a9", "network": {"id": "ddc08dc9-d08f-4fb9-9bee-1a030ed69f76", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1124067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b33dea0794a48f78b9f519cb269a8c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f1b07b1-e4e5-4842-9090-07fb2c3e124b", "external-id": "nsx-vlan-transportzone-646", "segmentation_id": 646, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1af3a42f-34", "ovs_interfaceid": "1af3a42f-34db-4c40-8efc-330a284d1f2c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.041487] env[61906]: DEBUG nova.compute.manager [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 901.055634] env[61906]: DEBUG oslo_vmware.api [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333293, 'name': CloneVM_Task, 'duration_secs': 1.543035} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.055920] env[61906]: INFO nova.virt.vmwareapi.vmops [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Created linked-clone VM from snapshot [ 901.056977] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac1f5b62-b1d2-4ce1-8838-44e529c06dad {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.067909] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Uploading image 22249871-324b-4f3e-9a02-9f2cc7665edb {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 901.072491] env[61906]: DEBUG nova.virt.hardware [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 901.072773] env[61906]: DEBUG nova.virt.hardware [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 901.072943] env[61906]: DEBUG nova.virt.hardware [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 901.073146] env[61906]: DEBUG nova.virt.hardware [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 901.073300] env[61906]: DEBUG nova.virt.hardware [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 901.073453] env[61906]: DEBUG nova.virt.hardware [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 901.073668] env[61906]: DEBUG nova.virt.hardware [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 901.073847] env[61906]: DEBUG nova.virt.hardware [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 901.074035] env[61906]: DEBUG nova.virt.hardware [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 901.074212] env[61906]: DEBUG nova.virt.hardware [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 901.074395] env[61906]: DEBUG nova.virt.hardware [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 901.075623] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-989fe03c-da62-4ebd-af59-51d9ffedfe9e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.085064] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc903b3-3d13-449f-b487-790d40be682b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.104963] env[61906]: DEBUG oslo_vmware.rw_handles [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 901.104963] env[61906]: value = "vm-284774" [ 901.104963] env[61906]: _type = "VirtualMachine" [ 901.104963] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 901.105295] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-83d7371a-ce5a-44e9-8cd1-aa39c1979cad {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.116828] env[61906]: DEBUG oslo_vmware.rw_handles [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lease: (returnval){ [ 901.116828] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5248bd5c-173b-0e12-133d-c4fc2b5f582f" [ 901.116828] env[61906]: _type = "HttpNfcLease" [ 901.116828] env[61906]: } obtained for exporting VM: (result){ [ 901.116828] env[61906]: value = "vm-284774" [ 901.116828] env[61906]: _type = "VirtualMachine" [ 901.116828] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 901.117204] env[61906]: DEBUG oslo_vmware.api [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the lease: (returnval){ [ 901.117204] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5248bd5c-173b-0e12-133d-c4fc2b5f582f" [ 901.117204] env[61906]: _type = "HttpNfcLease" [ 901.117204] env[61906]: } to be ready. {{(pid=61906) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 901.127087] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 901.127087] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5248bd5c-173b-0e12-133d-c4fc2b5f582f" [ 901.127087] env[61906]: _type = "HttpNfcLease" [ 901.127087] env[61906]: } is initializing. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 901.150594] env[61906]: DEBUG oslo_vmware.api [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333298, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.179946] env[61906]: DEBUG oslo_vmware.api [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333297, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.193395] env[61906]: DEBUG nova.compute.manager [req-910391e2-4fe3-410b-aa79-1f18a35c2f3e req-441fbedc-44bf-475f-93ae-5d9feb5f6cfa service nova] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Received event network-changed-f76ea64e-15c4-490e-9a5e-4c244195d983 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 901.193395] env[61906]: DEBUG nova.compute.manager [req-910391e2-4fe3-410b-aa79-1f18a35c2f3e req-441fbedc-44bf-475f-93ae-5d9feb5f6cfa service nova] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Refreshing instance network info cache due to event network-changed-f76ea64e-15c4-490e-9a5e-4c244195d983. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 901.193578] env[61906]: DEBUG oslo_concurrency.lockutils [req-910391e2-4fe3-410b-aa79-1f18a35c2f3e req-441fbedc-44bf-475f-93ae-5d9feb5f6cfa service nova] Acquiring lock "refresh_cache-da493512-d996-4de7-9f47-cadcbc4fbcb9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.193621] env[61906]: DEBUG oslo_concurrency.lockutils [req-910391e2-4fe3-410b-aa79-1f18a35c2f3e req-441fbedc-44bf-475f-93ae-5d9feb5f6cfa service nova] Acquired lock "refresh_cache-da493512-d996-4de7-9f47-cadcbc4fbcb9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.193764] env[61906]: DEBUG nova.network.neutron [req-910391e2-4fe3-410b-aa79-1f18a35c2f3e req-441fbedc-44bf-475f-93ae-5d9feb5f6cfa service nova] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Refreshing network info cache for port f76ea64e-15c4-490e-9a5e-4c244195d983 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 901.326830] env[61906]: DEBUG nova.scheduler.client.report [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 901.349301] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Releasing lock "refresh_cache-1bc9f3d4-b539-4841-b5b6-eaefa0d06a72" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.349684] env[61906]: DEBUG nova.compute.manager [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Instance network_info: |[{"id": "1af3a42f-34db-4c40-8efc-330a284d1f2c", "address": "fa:16:3e:39:df:a9", "network": {"id": "ddc08dc9-d08f-4fb9-9bee-1a030ed69f76", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1124067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b33dea0794a48f78b9f519cb269a8c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f1b07b1-e4e5-4842-9090-07fb2c3e124b", "external-id": "nsx-vlan-transportzone-646", "segmentation_id": 646, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1af3a42f-34", "ovs_interfaceid": "1af3a42f-34db-4c40-8efc-330a284d1f2c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 901.350175] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:39:df:a9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6f1b07b1-e4e5-4842-9090-07fb2c3e124b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1af3a42f-34db-4c40-8efc-330a284d1f2c', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 901.363398] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Creating folder: Project (1b33dea0794a48f78b9f519cb269a8c2). Parent ref: group-v284713. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 901.363849] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6291b8ec-3b93-40eb-b340-0283d47118fd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.389471] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Created folder: Project (1b33dea0794a48f78b9f519cb269a8c2) in parent group-v284713. [ 901.389804] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Creating folder: Instances. Parent ref: group-v284779. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 901.389998] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b34f4391-2e4d-4d10-9b07-822264f42322 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.401748] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Created folder: Instances in parent group-v284779. [ 901.404041] env[61906]: DEBUG oslo.service.loopingcall [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.404041] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 901.404041] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c4ed1266-3978-433b-b3ac-84d69da94105 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.422689] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 901.422689] env[61906]: value = "task-1333302" [ 901.422689] env[61906]: _type = "Task" [ 901.422689] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.426568] env[61906]: DEBUG nova.network.neutron [req-f636d7bf-27a6-4f4b-997f-e3e0c158000f req-a5ab416c-3c5b-459a-947f-9c9c767ef98a service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Updated VIF entry in instance network info cache for port 8bcec3ac-5104-4e72-af50-08fae9dc156c. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 901.427018] env[61906]: DEBUG nova.network.neutron [req-f636d7bf-27a6-4f4b-997f-e3e0c158000f req-a5ab416c-3c5b-459a-947f-9c9c767ef98a service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Updating instance_info_cache with network_info: [{"id": "8bcec3ac-5104-4e72-af50-08fae9dc156c", "address": "fa:16:3e:36:f8:fd", "network": {"id": "eaf64f00-dd3b-41c3-b679-3ee2f53735a0", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-255429219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "5375ff2ae28b4427a3c4541ee3d2729a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20fbc86c-0935-406c-9659-3ffc33fc0d08", "external-id": "nsx-vlan-transportzone-191", "segmentation_id": 191, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bcec3ac-51", "ovs_interfaceid": "8bcec3ac-5104-4e72-af50-08fae9dc156c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.433961] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333302, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.629297] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 901.629297] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5248bd5c-173b-0e12-133d-c4fc2b5f582f" [ 901.629297] env[61906]: _type = "HttpNfcLease" [ 901.629297] env[61906]: } is ready. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 901.629661] env[61906]: DEBUG oslo_vmware.rw_handles [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 901.629661] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5248bd5c-173b-0e12-133d-c4fc2b5f582f" [ 901.629661] env[61906]: _type = "HttpNfcLease" [ 901.629661] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 901.632218] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f9ff7a9-d5f5-4d48-8826-0850856749d6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.640961] env[61906]: DEBUG oslo_vmware.rw_handles [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524b4823-46bc-bfd8-bb2d-bc2ecac8f1ee/disk-0.vmdk from lease info. {{(pid=61906) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 901.640961] env[61906]: DEBUG oslo_vmware.rw_handles [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524b4823-46bc-bfd8-bb2d-bc2ecac8f1ee/disk-0.vmdk for reading. {{(pid=61906) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 901.718791] env[61906]: DEBUG oslo_vmware.api [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333298, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.734087} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.720330] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] b1159533-c970-49d6-af42-b954b20d92fb/b1159533-c970-49d6-af42-b954b20d92fb.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 901.720561] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 901.721135] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-989d6d0f-077c-46d2-ae00-eb6c2661c9ec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.726818] env[61906]: DEBUG oslo_vmware.api [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333297, 'name': CreateSnapshot_Task, 'duration_secs': 1.236735} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.727105] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Created Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 901.727912] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d6108f4-a504-4183-b906-9adb6e42eca7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.740462] env[61906]: DEBUG oslo_vmware.api [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 901.740462] env[61906]: value = "task-1333303" [ 901.740462] env[61906]: _type = "Task" [ 901.740462] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.750142] env[61906]: DEBUG oslo_vmware.api [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333303, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.769327] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-bc4bc64a-5dc9-470f-9b76-4fb6848c2b7d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.833837] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.810s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.838028] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.883s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.838821] env[61906]: INFO nova.compute.claims [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 901.861943] env[61906]: INFO nova.scheduler.client.report [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Deleted allocations for instance 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8 [ 901.929878] env[61906]: DEBUG oslo_concurrency.lockutils [req-f636d7bf-27a6-4f4b-997f-e3e0c158000f req-a5ab416c-3c5b-459a-947f-9c9c767ef98a service nova] Releasing lock "refresh_cache-75b9738e-e9b6-435a-88bb-851982828d36" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.930181] env[61906]: DEBUG nova.compute.manager [req-f636d7bf-27a6-4f4b-997f-e3e0c158000f req-a5ab416c-3c5b-459a-947f-9c9c767ef98a service nova] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Received event network-vif-plugged-1af3a42f-34db-4c40-8efc-330a284d1f2c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 901.930964] env[61906]: DEBUG oslo_concurrency.lockutils [req-f636d7bf-27a6-4f4b-997f-e3e0c158000f req-a5ab416c-3c5b-459a-947f-9c9c767ef98a service nova] Acquiring lock "1bc9f3d4-b539-4841-b5b6-eaefa0d06a72-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.930964] env[61906]: DEBUG oslo_concurrency.lockutils [req-f636d7bf-27a6-4f4b-997f-e3e0c158000f req-a5ab416c-3c5b-459a-947f-9c9c767ef98a service nova] Lock "1bc9f3d4-b539-4841-b5b6-eaefa0d06a72-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.930964] env[61906]: DEBUG oslo_concurrency.lockutils [req-f636d7bf-27a6-4f4b-997f-e3e0c158000f req-a5ab416c-3c5b-459a-947f-9c9c767ef98a service nova] Lock "1bc9f3d4-b539-4841-b5b6-eaefa0d06a72-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.930964] env[61906]: DEBUG nova.compute.manager [req-f636d7bf-27a6-4f4b-997f-e3e0c158000f req-a5ab416c-3c5b-459a-947f-9c9c767ef98a service nova] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] No waiting events found dispatching network-vif-plugged-1af3a42f-34db-4c40-8efc-330a284d1f2c {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 901.931371] env[61906]: WARNING nova.compute.manager [req-f636d7bf-27a6-4f4b-997f-e3e0c158000f req-a5ab416c-3c5b-459a-947f-9c9c767ef98a service nova] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Received unexpected event network-vif-plugged-1af3a42f-34db-4c40-8efc-330a284d1f2c for instance with vm_state building and task_state spawning. [ 901.931371] env[61906]: DEBUG nova.compute.manager [req-f636d7bf-27a6-4f4b-997f-e3e0c158000f req-a5ab416c-3c5b-459a-947f-9c9c767ef98a service nova] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Received event network-changed-1af3a42f-34db-4c40-8efc-330a284d1f2c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 901.931486] env[61906]: DEBUG nova.compute.manager [req-f636d7bf-27a6-4f4b-997f-e3e0c158000f req-a5ab416c-3c5b-459a-947f-9c9c767ef98a service nova] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Refreshing instance network info cache due to event network-changed-1af3a42f-34db-4c40-8efc-330a284d1f2c. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 901.931701] env[61906]: DEBUG oslo_concurrency.lockutils [req-f636d7bf-27a6-4f4b-997f-e3e0c158000f req-a5ab416c-3c5b-459a-947f-9c9c767ef98a service nova] Acquiring lock "refresh_cache-1bc9f3d4-b539-4841-b5b6-eaefa0d06a72" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.931845] env[61906]: DEBUG oslo_concurrency.lockutils [req-f636d7bf-27a6-4f4b-997f-e3e0c158000f req-a5ab416c-3c5b-459a-947f-9c9c767ef98a service nova] Acquired lock "refresh_cache-1bc9f3d4-b539-4841-b5b6-eaefa0d06a72" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.932012] env[61906]: DEBUG nova.network.neutron [req-f636d7bf-27a6-4f4b-997f-e3e0c158000f req-a5ab416c-3c5b-459a-947f-9c9c767ef98a service nova] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Refreshing network info cache for port 1af3a42f-34db-4c40-8efc-330a284d1f2c {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 901.940537] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333302, 'name': CreateVM_Task, 'duration_secs': 0.43238} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.940537] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 901.940797] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.940797] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.941210] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 901.941893] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-293fcbaf-322d-4306-b90d-6504870bcda8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.947831] env[61906]: DEBUG oslo_vmware.api [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 901.947831] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]525dc3bc-61e4-0e1b-3f94-8d1fc371c472" [ 901.947831] env[61906]: _type = "Task" [ 901.947831] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.956705] env[61906]: DEBUG oslo_vmware.api [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]525dc3bc-61e4-0e1b-3f94-8d1fc371c472, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.081445] env[61906]: DEBUG nova.network.neutron [req-910391e2-4fe3-410b-aa79-1f18a35c2f3e req-441fbedc-44bf-475f-93ae-5d9feb5f6cfa service nova] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Updated VIF entry in instance network info cache for port f76ea64e-15c4-490e-9a5e-4c244195d983. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 902.081716] env[61906]: DEBUG nova.network.neutron [req-910391e2-4fe3-410b-aa79-1f18a35c2f3e req-441fbedc-44bf-475f-93ae-5d9feb5f6cfa service nova] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Updating instance_info_cache with network_info: [{"id": "f76ea64e-15c4-490e-9a5e-4c244195d983", "address": "fa:16:3e:68:ee:08", "network": {"id": "b5a4d04c-7682-4981-9e34-c416774c62a0", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-554429652-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "527d3de1bab74c4a8684edba72b6cb3f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c979f78-8597-41f8-b1de-995014032689", "external-id": "nsx-vlan-transportzone-477", "segmentation_id": 477, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf76ea64e-15", "ovs_interfaceid": "f76ea64e-15c4-490e-9a5e-4c244195d983", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.124228] env[61906]: DEBUG nova.network.neutron [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Successfully updated port: 7e726302-5576-4a98-b2a5-23c4ed0be21e {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 902.250062] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Creating linked-clone VM from snapshot {{(pid=61906) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 902.250574] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-c60c2852-6b75-4e8c-9d30-6ba43ffac010 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.263929] env[61906]: DEBUG oslo_vmware.api [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333303, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.100355} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.265948] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 902.266141] env[61906]: DEBUG oslo_vmware.api [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Waiting for the task: (returnval){ [ 902.266141] env[61906]: value = "task-1333304" [ 902.266141] env[61906]: _type = "Task" [ 902.266141] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.267044] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0010a64e-842f-410c-82c2-50d1e42b9679 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.296512] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] b1159533-c970-49d6-af42-b954b20d92fb/b1159533-c970-49d6-af42-b954b20d92fb.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 902.300753] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a70f80bc-3c4c-4a42-93c0-1e14a733ab68 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.317771] env[61906]: DEBUG oslo_vmware.api [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333304, 'name': CloneVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.324970] env[61906]: DEBUG oslo_vmware.api [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 902.324970] env[61906]: value = "task-1333305" [ 902.324970] env[61906]: _type = "Task" [ 902.324970] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.334220] env[61906]: DEBUG oslo_vmware.api [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333305, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.372165] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d1deb2a3-687c-4adf-9562-ef18b0b17ab8 tempest-ServersAaction247Test-417589248 tempest-ServersAaction247Test-417589248-project-member] Lock "3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.456s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.460454] env[61906]: DEBUG oslo_vmware.api [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]525dc3bc-61e4-0e1b-3f94-8d1fc371c472, 'name': SearchDatastore_Task, 'duration_secs': 0.038856} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.461277] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.462010] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 902.462010] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.462202] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.462512] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 902.462850] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b4e31d23-1e0f-4662-916b-10d75b9a17a6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.477205] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 902.477476] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 902.478263] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d777c05-a473-48d5-9abd-43e0fe616f4b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.486658] env[61906]: DEBUG oslo_vmware.api [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 902.486658] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]525245f7-add6-db12-209f-6aa8b4ee06e5" [ 902.486658] env[61906]: _type = "Task" [ 902.486658] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.498316] env[61906]: DEBUG oslo_vmware.api [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]525245f7-add6-db12-209f-6aa8b4ee06e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.587837] env[61906]: DEBUG oslo_concurrency.lockutils [req-910391e2-4fe3-410b-aa79-1f18a35c2f3e req-441fbedc-44bf-475f-93ae-5d9feb5f6cfa service nova] Releasing lock "refresh_cache-da493512-d996-4de7-9f47-cadcbc4fbcb9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.588368] env[61906]: DEBUG nova.compute.manager [req-910391e2-4fe3-410b-aa79-1f18a35c2f3e req-441fbedc-44bf-475f-93ae-5d9feb5f6cfa service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Received event network-changed-8bcec3ac-5104-4e72-af50-08fae9dc156c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.589265] env[61906]: DEBUG nova.compute.manager [req-910391e2-4fe3-410b-aa79-1f18a35c2f3e req-441fbedc-44bf-475f-93ae-5d9feb5f6cfa service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Refreshing instance network info cache due to event network-changed-8bcec3ac-5104-4e72-af50-08fae9dc156c. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 902.589265] env[61906]: DEBUG oslo_concurrency.lockutils [req-910391e2-4fe3-410b-aa79-1f18a35c2f3e req-441fbedc-44bf-475f-93ae-5d9feb5f6cfa service nova] Acquiring lock "refresh_cache-75b9738e-e9b6-435a-88bb-851982828d36" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.589265] env[61906]: DEBUG oslo_concurrency.lockutils [req-910391e2-4fe3-410b-aa79-1f18a35c2f3e req-441fbedc-44bf-475f-93ae-5d9feb5f6cfa service nova] Acquired lock "refresh_cache-75b9738e-e9b6-435a-88bb-851982828d36" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.589265] env[61906]: DEBUG nova.network.neutron [req-910391e2-4fe3-410b-aa79-1f18a35c2f3e req-441fbedc-44bf-475f-93ae-5d9feb5f6cfa service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Refreshing network info cache for port 8bcec3ac-5104-4e72-af50-08fae9dc156c {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 902.627699] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Acquiring lock "refresh_cache-0e86bfdf-9002-4a2d-a2ac-af35921b020e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.627814] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Acquired lock "refresh_cache-0e86bfdf-9002-4a2d-a2ac-af35921b020e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.627982] env[61906]: DEBUG nova.network.neutron [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 902.785601] env[61906]: DEBUG oslo_vmware.api [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333304, 'name': CloneVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.795833] env[61906]: DEBUG nova.network.neutron [req-f636d7bf-27a6-4f4b-997f-e3e0c158000f req-a5ab416c-3c5b-459a-947f-9c9c767ef98a service nova] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Updated VIF entry in instance network info cache for port 1af3a42f-34db-4c40-8efc-330a284d1f2c. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 902.796284] env[61906]: DEBUG nova.network.neutron [req-f636d7bf-27a6-4f4b-997f-e3e0c158000f req-a5ab416c-3c5b-459a-947f-9c9c767ef98a service nova] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Updating instance_info_cache with network_info: [{"id": "1af3a42f-34db-4c40-8efc-330a284d1f2c", "address": "fa:16:3e:39:df:a9", "network": {"id": "ddc08dc9-d08f-4fb9-9bee-1a030ed69f76", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1124067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b33dea0794a48f78b9f519cb269a8c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f1b07b1-e4e5-4842-9090-07fb2c3e124b", "external-id": "nsx-vlan-transportzone-646", "segmentation_id": 646, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1af3a42f-34", "ovs_interfaceid": "1af3a42f-34db-4c40-8efc-330a284d1f2c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.839077] env[61906]: DEBUG oslo_vmware.api [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333305, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.009892] env[61906]: DEBUG oslo_vmware.api [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]525245f7-add6-db12-209f-6aa8b4ee06e5, 'name': SearchDatastore_Task, 'duration_secs': 0.01819} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.010905] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18754975-4499-420e-9f19-c8d47c0092dc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.021465] env[61906]: DEBUG oslo_vmware.api [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 903.021465] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5275da1d-a88d-063c-39e3-c09db839b636" [ 903.021465] env[61906]: _type = "Task" [ 903.021465] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.048229] env[61906]: DEBUG oslo_vmware.api [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5275da1d-a88d-063c-39e3-c09db839b636, 'name': SearchDatastore_Task, 'duration_secs': 0.024121} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.050318] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.050782] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72/1bc9f3d4-b539-4841-b5b6-eaefa0d06a72.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 903.052251] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f23e007-4943-40d2-9272-8504fe338289 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.059276] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fc5aa591-1b9c-4958-9112-bf232fa2963f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.077518] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e305106-833b-434f-b93a-df35fd197847 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.082953] env[61906]: DEBUG oslo_vmware.api [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 903.082953] env[61906]: value = "task-1333306" [ 903.082953] env[61906]: _type = "Task" [ 903.082953] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.121106] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72f9607c-d0d3-4bc8-bfe4-eb64f2a68eb9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.130569] env[61906]: DEBUG oslo_vmware.api [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333306, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.138278] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13ba3b2c-0c56-4409-a30f-5182e6ce1476 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.159260] env[61906]: DEBUG nova.compute.provider_tree [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.188899] env[61906]: DEBUG nova.network.neutron [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 903.230537] env[61906]: DEBUG nova.compute.manager [req-9c3edf2d-2ba6-4430-9646-320ff0785d23 req-e8e57ee1-8a92-4514-94c2-a8e95da13355 service nova] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Received event network-vif-plugged-7e726302-5576-4a98-b2a5-23c4ed0be21e {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 903.230615] env[61906]: DEBUG oslo_concurrency.lockutils [req-9c3edf2d-2ba6-4430-9646-320ff0785d23 req-e8e57ee1-8a92-4514-94c2-a8e95da13355 service nova] Acquiring lock "0e86bfdf-9002-4a2d-a2ac-af35921b020e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.230806] env[61906]: DEBUG oslo_concurrency.lockutils [req-9c3edf2d-2ba6-4430-9646-320ff0785d23 req-e8e57ee1-8a92-4514-94c2-a8e95da13355 service nova] Lock "0e86bfdf-9002-4a2d-a2ac-af35921b020e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.230979] env[61906]: DEBUG oslo_concurrency.lockutils [req-9c3edf2d-2ba6-4430-9646-320ff0785d23 req-e8e57ee1-8a92-4514-94c2-a8e95da13355 service nova] Lock "0e86bfdf-9002-4a2d-a2ac-af35921b020e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.231154] env[61906]: DEBUG nova.compute.manager [req-9c3edf2d-2ba6-4430-9646-320ff0785d23 req-e8e57ee1-8a92-4514-94c2-a8e95da13355 service nova] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] No waiting events found dispatching network-vif-plugged-7e726302-5576-4a98-b2a5-23c4ed0be21e {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 903.231321] env[61906]: WARNING nova.compute.manager [req-9c3edf2d-2ba6-4430-9646-320ff0785d23 req-e8e57ee1-8a92-4514-94c2-a8e95da13355 service nova] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Received unexpected event network-vif-plugged-7e726302-5576-4a98-b2a5-23c4ed0be21e for instance with vm_state building and task_state spawning. [ 903.231482] env[61906]: DEBUG nova.compute.manager [req-9c3edf2d-2ba6-4430-9646-320ff0785d23 req-e8e57ee1-8a92-4514-94c2-a8e95da13355 service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Received event network-changed-8bcec3ac-5104-4e72-af50-08fae9dc156c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 903.231634] env[61906]: DEBUG nova.compute.manager [req-9c3edf2d-2ba6-4430-9646-320ff0785d23 req-e8e57ee1-8a92-4514-94c2-a8e95da13355 service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Refreshing instance network info cache due to event network-changed-8bcec3ac-5104-4e72-af50-08fae9dc156c. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 903.231799] env[61906]: DEBUG oslo_concurrency.lockutils [req-9c3edf2d-2ba6-4430-9646-320ff0785d23 req-e8e57ee1-8a92-4514-94c2-a8e95da13355 service nova] Acquiring lock "refresh_cache-75b9738e-e9b6-435a-88bb-851982828d36" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.283652] env[61906]: DEBUG oslo_vmware.api [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333304, 'name': CloneVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.303591] env[61906]: DEBUG oslo_concurrency.lockutils [req-f636d7bf-27a6-4f4b-997f-e3e0c158000f req-a5ab416c-3c5b-459a-947f-9c9c767ef98a service nova] Releasing lock "refresh_cache-1bc9f3d4-b539-4841-b5b6-eaefa0d06a72" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.338306] env[61906]: DEBUG oslo_vmware.api [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333305, 'name': ReconfigVM_Task, 'duration_secs': 0.75761} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.338714] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Reconfigured VM instance instance-00000048 to attach disk [datastore2] b1159533-c970-49d6-af42-b954b20d92fb/b1159533-c970-49d6-af42-b954b20d92fb.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 903.339340] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9d020eea-4bd2-476e-984a-d78ab7f3100f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.351492] env[61906]: DEBUG oslo_vmware.api [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 903.351492] env[61906]: value = "task-1333307" [ 903.351492] env[61906]: _type = "Task" [ 903.351492] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.363646] env[61906]: DEBUG oslo_vmware.api [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333307, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.433600] env[61906]: DEBUG nova.network.neutron [req-910391e2-4fe3-410b-aa79-1f18a35c2f3e req-441fbedc-44bf-475f-93ae-5d9feb5f6cfa service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Updated VIF entry in instance network info cache for port 8bcec3ac-5104-4e72-af50-08fae9dc156c. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 903.434017] env[61906]: DEBUG nova.network.neutron [req-910391e2-4fe3-410b-aa79-1f18a35c2f3e req-441fbedc-44bf-475f-93ae-5d9feb5f6cfa service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Updating instance_info_cache with network_info: [{"id": "8bcec3ac-5104-4e72-af50-08fae9dc156c", "address": "fa:16:3e:36:f8:fd", "network": {"id": "eaf64f00-dd3b-41c3-b679-3ee2f53735a0", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-255429219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "5375ff2ae28b4427a3c4541ee3d2729a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20fbc86c-0935-406c-9659-3ffc33fc0d08", "external-id": "nsx-vlan-transportzone-191", "segmentation_id": 191, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bcec3ac-51", "ovs_interfaceid": "8bcec3ac-5104-4e72-af50-08fae9dc156c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.535903] env[61906]: DEBUG nova.network.neutron [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Updating instance_info_cache with network_info: [{"id": "7e726302-5576-4a98-b2a5-23c4ed0be21e", "address": "fa:16:3e:e0:f7:dd", "network": {"id": "2240e54b-fc37-4d48-9ceb-a386504ab2c2", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-771370996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "357662e4b2e7430b82a521ed0f4392ad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5bc2dd1b-c901-4f3d-9c3a-af171c275c4b", "external-id": "nsx-vlan-transportzone-320", "segmentation_id": 320, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e726302-55", "ovs_interfaceid": "7e726302-5576-4a98-b2a5-23c4ed0be21e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.596027] env[61906]: DEBUG oslo_vmware.api [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333306, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.631688] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Acquiring lock "75b9738e-e9b6-435a-88bb-851982828d36" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.632259] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Lock "75b9738e-e9b6-435a-88bb-851982828d36" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.632552] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Acquiring lock "75b9738e-e9b6-435a-88bb-851982828d36-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.632791] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Lock "75b9738e-e9b6-435a-88bb-851982828d36-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.632979] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Lock "75b9738e-e9b6-435a-88bb-851982828d36-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.635582] env[61906]: INFO nova.compute.manager [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Terminating instance [ 903.637744] env[61906]: DEBUG nova.compute.manager [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 903.637952] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 903.638852] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cdfd03b-f79f-46de-9dac-fa57ed0300b5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.647504] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 903.647838] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-545d13ca-e6b8-43bb-859d-789cfd741b18 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.656257] env[61906]: DEBUG oslo_vmware.api [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Waiting for the task: (returnval){ [ 903.656257] env[61906]: value = "task-1333308" [ 903.656257] env[61906]: _type = "Task" [ 903.656257] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.664108] env[61906]: DEBUG nova.scheduler.client.report [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 903.672984] env[61906]: DEBUG oslo_vmware.api [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333308, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.785590] env[61906]: DEBUG oslo_vmware.api [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333304, 'name': CloneVM_Task} progress is 95%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.863777] env[61906]: DEBUG oslo_vmware.api [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333307, 'name': Rename_Task, 'duration_secs': 0.375} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.864273] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 903.864531] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9b3fd6c0-ad3c-4e11-8772-ef730f23ea00 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.875198] env[61906]: DEBUG oslo_vmware.api [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 903.875198] env[61906]: value = "task-1333309" [ 903.875198] env[61906]: _type = "Task" [ 903.875198] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.883597] env[61906]: DEBUG oslo_vmware.api [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333309, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.937135] env[61906]: DEBUG oslo_concurrency.lockutils [req-910391e2-4fe3-410b-aa79-1f18a35c2f3e req-441fbedc-44bf-475f-93ae-5d9feb5f6cfa service nova] Releasing lock "refresh_cache-75b9738e-e9b6-435a-88bb-851982828d36" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.937647] env[61906]: DEBUG oslo_concurrency.lockutils [req-9c3edf2d-2ba6-4430-9646-320ff0785d23 req-e8e57ee1-8a92-4514-94c2-a8e95da13355 service nova] Acquired lock "refresh_cache-75b9738e-e9b6-435a-88bb-851982828d36" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.937849] env[61906]: DEBUG nova.network.neutron [req-9c3edf2d-2ba6-4430-9646-320ff0785d23 req-e8e57ee1-8a92-4514-94c2-a8e95da13355 service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Refreshing network info cache for port 8bcec3ac-5104-4e72-af50-08fae9dc156c {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 904.040175] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Releasing lock "refresh_cache-0e86bfdf-9002-4a2d-a2ac-af35921b020e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.040599] env[61906]: DEBUG nova.compute.manager [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Instance network_info: |[{"id": "7e726302-5576-4a98-b2a5-23c4ed0be21e", "address": "fa:16:3e:e0:f7:dd", "network": {"id": "2240e54b-fc37-4d48-9ceb-a386504ab2c2", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-771370996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "357662e4b2e7430b82a521ed0f4392ad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5bc2dd1b-c901-4f3d-9c3a-af171c275c4b", "external-id": "nsx-vlan-transportzone-320", "segmentation_id": 320, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e726302-55", "ovs_interfaceid": "7e726302-5576-4a98-b2a5-23c4ed0be21e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 904.041095] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e0:f7:dd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5bc2dd1b-c901-4f3d-9c3a-af171c275c4b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7e726302-5576-4a98-b2a5-23c4ed0be21e', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 904.049427] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Creating folder: Project (357662e4b2e7430b82a521ed0f4392ad). Parent ref: group-v284713. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 904.049746] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-996a542a-f77b-4c53-aff5-22e62e526c27 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.064572] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Created folder: Project (357662e4b2e7430b82a521ed0f4392ad) in parent group-v284713. [ 904.064799] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Creating folder: Instances. Parent ref: group-v284783. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 904.065354] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2126af19-3753-4403-9d1a-aac7f86e1a8a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.075334] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Created folder: Instances in parent group-v284783. [ 904.075682] env[61906]: DEBUG oslo.service.loopingcall [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 904.075822] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 904.076054] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-414160ad-344b-4d27-ac7c-a50d34ed7729 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.101717] env[61906]: DEBUG oslo_vmware.api [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333306, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.559451} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.103823] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72/1bc9f3d4-b539-4841-b5b6-eaefa0d06a72.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 904.104067] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 904.104316] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 904.104316] env[61906]: value = "task-1333312" [ 904.104316] env[61906]: _type = "Task" [ 904.104316] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.104717] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8384f28a-5e77-4cd5-86d4-9c041af33e30 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.117363] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333312, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.118313] env[61906]: DEBUG oslo_vmware.api [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 904.118313] env[61906]: value = "task-1333313" [ 904.118313] env[61906]: _type = "Task" [ 904.118313] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.168310] env[61906]: DEBUG oslo_vmware.api [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333308, 'name': PowerOffVM_Task, 'duration_secs': 0.233402} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.168676] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 904.168884] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 904.169895] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5174985c-042f-4d71-aa1c-ac81a7debcc1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.174847] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.338s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.175449] env[61906]: DEBUG nova.compute.manager [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 904.178447] env[61906]: DEBUG oslo_concurrency.lockutils [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.673s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.178629] env[61906]: DEBUG nova.objects.instance [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lazy-loading 'resources' on Instance uuid 9f1253e6-4727-4dca-bad2-04f0c5424cdb {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 904.245364] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 904.245984] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 904.245984] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Deleting the datastore file [datastore1] 75b9738e-e9b6-435a-88bb-851982828d36 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 904.246120] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a382f478-fd21-4183-bddd-90d8401c2557 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.253537] env[61906]: DEBUG oslo_vmware.api [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Waiting for the task: (returnval){ [ 904.253537] env[61906]: value = "task-1333315" [ 904.253537] env[61906]: _type = "Task" [ 904.253537] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.263834] env[61906]: DEBUG oslo_vmware.api [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333315, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.281979] env[61906]: DEBUG oslo_vmware.api [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333304, 'name': CloneVM_Task, 'duration_secs': 1.731876} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.282410] env[61906]: INFO nova.virt.vmwareapi.vmops [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Created linked-clone VM from snapshot [ 904.283240] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1e0c174-1036-48b4-ad32-7978af3fac5a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.292893] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Uploading image 4efc4cce-9185-4c01-8fa1-4d7a2414b871 {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 904.320057] env[61906]: DEBUG oslo_vmware.rw_handles [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 904.320057] env[61906]: value = "vm-284782" [ 904.320057] env[61906]: _type = "VirtualMachine" [ 904.320057] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 904.320474] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-7c2e0243-39ad-4492-8290-1eb8d1855d2e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.331438] env[61906]: DEBUG oslo_vmware.rw_handles [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Lease: (returnval){ [ 904.331438] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]523cc058-f1ec-9938-52eb-a6705ccbed3d" [ 904.331438] env[61906]: _type = "HttpNfcLease" [ 904.331438] env[61906]: } obtained for exporting VM: (result){ [ 904.331438] env[61906]: value = "vm-284782" [ 904.331438] env[61906]: _type = "VirtualMachine" [ 904.331438] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 904.331863] env[61906]: DEBUG oslo_vmware.api [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Waiting for the lease: (returnval){ [ 904.331863] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]523cc058-f1ec-9938-52eb-a6705ccbed3d" [ 904.331863] env[61906]: _type = "HttpNfcLease" [ 904.331863] env[61906]: } to be ready. {{(pid=61906) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 904.340498] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 904.340498] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]523cc058-f1ec-9938-52eb-a6705ccbed3d" [ 904.340498] env[61906]: _type = "HttpNfcLease" [ 904.340498] env[61906]: } is initializing. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 904.388462] env[61906]: DEBUG oslo_vmware.api [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333309, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.618390] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333312, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.628253] env[61906]: DEBUG oslo_vmware.api [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333313, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089708} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.628638] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 904.629572] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db39c05d-4edb-4b09-959a-6d8f0225cccf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.657855] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72/1bc9f3d4-b539-4841-b5b6-eaefa0d06a72.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 904.658286] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fb080aaf-d6e6-4ff3-b181-685e6f0e54a7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.686025] env[61906]: DEBUG nova.compute.utils [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 904.690356] env[61906]: DEBUG nova.compute.manager [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 904.690926] env[61906]: DEBUG nova.network.neutron [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 904.693197] env[61906]: DEBUG oslo_vmware.api [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 904.693197] env[61906]: value = "task-1333317" [ 904.693197] env[61906]: _type = "Task" [ 904.693197] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.705428] env[61906]: DEBUG oslo_vmware.api [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333317, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.769172] env[61906]: DEBUG oslo_vmware.api [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Task: {'id': task-1333315, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.343503} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.771116] env[61906]: DEBUG nova.policy [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bfb8ebd8f94f45dfadb1d8802d04aec9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4f7fc2042e6549cabe98838990f35842', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 904.773172] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 904.773673] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 904.775652] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 904.775652] env[61906]: INFO nova.compute.manager [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Took 1.14 seconds to destroy the instance on the hypervisor. [ 904.775652] env[61906]: DEBUG oslo.service.loopingcall [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 904.779306] env[61906]: DEBUG nova.compute.manager [-] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 904.779306] env[61906]: DEBUG nova.network.neutron [-] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 904.846226] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 904.846226] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]523cc058-f1ec-9938-52eb-a6705ccbed3d" [ 904.846226] env[61906]: _type = "HttpNfcLease" [ 904.846226] env[61906]: } is ready. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 904.846879] env[61906]: DEBUG oslo_vmware.rw_handles [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 904.846879] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]523cc058-f1ec-9938-52eb-a6705ccbed3d" [ 904.846879] env[61906]: _type = "HttpNfcLease" [ 904.846879] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 904.848263] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c53eff56-7959-4eb3-8541-0250c75c2ac2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.863218] env[61906]: DEBUG oslo_vmware.rw_handles [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523d516e-d0c1-6da1-8a04-13c0867afd52/disk-0.vmdk from lease info. {{(pid=61906) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 904.863218] env[61906]: DEBUG oslo_vmware.rw_handles [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523d516e-d0c1-6da1-8a04-13c0867afd52/disk-0.vmdk for reading. {{(pid=61906) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 904.949023] env[61906]: DEBUG oslo_vmware.api [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333309, 'name': PowerOnVM_Task, 'duration_secs': 0.513115} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.949023] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 904.949023] env[61906]: INFO nova.compute.manager [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Took 8.64 seconds to spawn the instance on the hypervisor. [ 904.949023] env[61906]: DEBUG nova.compute.manager [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 904.949023] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96951e4b-00d9-4550-987d-17e9d39619f2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.951788] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbdbba59-d24b-4939-8bbd-1ee3913448f4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.965635] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19685226-f659-4676-aa2f-dd8e63893c56 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.998710] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-174c57c9-7e7c-4891-bb77-67a2e39026b0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.004641] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6728b546-2fd2-47a7-926b-7df26ecf4d9e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.011420] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d7a8c41-2f99-412b-84a4-6a52b66ca806 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.028630] env[61906]: DEBUG nova.compute.provider_tree [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.118260] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333312, 'name': CreateVM_Task, 'duration_secs': 0.61322} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.118451] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 905.119172] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.119347] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.119681] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 905.119944] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ceefeb9-37b0-4d6b-964b-5aaae6dbd90f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.126080] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Waiting for the task: (returnval){ [ 905.126080] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]528e0b8a-2905-d6d7-1539-a156003bd61c" [ 905.126080] env[61906]: _type = "Task" [ 905.126080] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.137761] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]528e0b8a-2905-d6d7-1539-a156003bd61c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.137761] env[61906]: DEBUG nova.network.neutron [req-9c3edf2d-2ba6-4430-9646-320ff0785d23 req-e8e57ee1-8a92-4514-94c2-a8e95da13355 service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Updated VIF entry in instance network info cache for port 8bcec3ac-5104-4e72-af50-08fae9dc156c. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 905.137761] env[61906]: DEBUG nova.network.neutron [req-9c3edf2d-2ba6-4430-9646-320ff0785d23 req-e8e57ee1-8a92-4514-94c2-a8e95da13355 service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Updating instance_info_cache with network_info: [{"id": "8bcec3ac-5104-4e72-af50-08fae9dc156c", "address": "fa:16:3e:36:f8:fd", "network": {"id": "eaf64f00-dd3b-41c3-b679-3ee2f53735a0", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-255429219-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "5375ff2ae28b4427a3c4541ee3d2729a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20fbc86c-0935-406c-9659-3ffc33fc0d08", "external-id": "nsx-vlan-transportzone-191", "segmentation_id": 191, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bcec3ac-51", "ovs_interfaceid": "8bcec3ac-5104-4e72-af50-08fae9dc156c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.169794] env[61906]: DEBUG nova.network.neutron [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Successfully created port: b54c27f8-7727-4166-af82-5b6fe0d3519e {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 905.189970] env[61906]: DEBUG nova.compute.manager [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 905.208973] env[61906]: DEBUG oslo_vmware.api [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333317, 'name': ReconfigVM_Task, 'duration_secs': 0.396231} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.209827] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72/1bc9f3d4-b539-4841-b5b6-eaefa0d06a72.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 905.210354] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-720e318b-2500-4823-beed-08c5b915a529 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.218636] env[61906]: DEBUG oslo_vmware.api [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 905.218636] env[61906]: value = "task-1333318" [ 905.218636] env[61906]: _type = "Task" [ 905.218636] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.232301] env[61906]: DEBUG oslo_vmware.api [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333318, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.278137] env[61906]: DEBUG nova.compute.manager [req-8e1bef8a-d13b-461e-b82a-8abb2ca57911 req-8a097445-5083-4471-94bc-a11f3d8f4f2f service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Received event network-vif-deleted-8bcec3ac-5104-4e72-af50-08fae9dc156c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 905.278409] env[61906]: INFO nova.compute.manager [req-8e1bef8a-d13b-461e-b82a-8abb2ca57911 req-8a097445-5083-4471-94bc-a11f3d8f4f2f service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Neutron deleted interface 8bcec3ac-5104-4e72-af50-08fae9dc156c; detaching it from the instance and deleting it from the info cache [ 905.278608] env[61906]: DEBUG nova.network.neutron [req-8e1bef8a-d13b-461e-b82a-8abb2ca57911 req-8a097445-5083-4471-94bc-a11f3d8f4f2f service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.476134] env[61906]: INFO nova.compute.manager [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Took 28.97 seconds to build instance. [ 905.536143] env[61906]: DEBUG nova.scheduler.client.report [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 905.640217] env[61906]: DEBUG oslo_concurrency.lockutils [req-9c3edf2d-2ba6-4430-9646-320ff0785d23 req-e8e57ee1-8a92-4514-94c2-a8e95da13355 service nova] Releasing lock "refresh_cache-75b9738e-e9b6-435a-88bb-851982828d36" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.640217] env[61906]: DEBUG nova.compute.manager [req-9c3edf2d-2ba6-4430-9646-320ff0785d23 req-e8e57ee1-8a92-4514-94c2-a8e95da13355 service nova] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Received event network-changed-7e726302-5576-4a98-b2a5-23c4ed0be21e {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 905.640661] env[61906]: DEBUG nova.compute.manager [req-9c3edf2d-2ba6-4430-9646-320ff0785d23 req-e8e57ee1-8a92-4514-94c2-a8e95da13355 service nova] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Refreshing instance network info cache due to event network-changed-7e726302-5576-4a98-b2a5-23c4ed0be21e. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 905.640948] env[61906]: DEBUG oslo_concurrency.lockutils [req-9c3edf2d-2ba6-4430-9646-320ff0785d23 req-e8e57ee1-8a92-4514-94c2-a8e95da13355 service nova] Acquiring lock "refresh_cache-0e86bfdf-9002-4a2d-a2ac-af35921b020e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.641253] env[61906]: DEBUG oslo_concurrency.lockutils [req-9c3edf2d-2ba6-4430-9646-320ff0785d23 req-e8e57ee1-8a92-4514-94c2-a8e95da13355 service nova] Acquired lock "refresh_cache-0e86bfdf-9002-4a2d-a2ac-af35921b020e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.641695] env[61906]: DEBUG nova.network.neutron [req-9c3edf2d-2ba6-4430-9646-320ff0785d23 req-e8e57ee1-8a92-4514-94c2-a8e95da13355 service nova] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Refreshing network info cache for port 7e726302-5576-4a98-b2a5-23c4ed0be21e {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 905.643715] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]528e0b8a-2905-d6d7-1539-a156003bd61c, 'name': SearchDatastore_Task, 'duration_secs': 0.017609} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.646703] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.646948] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 905.647324] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.647932] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.647932] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 905.648068] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fa123bfa-495f-43e0-aec4-d8ee39532fa8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.661731] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 905.662091] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 905.663680] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d239bc05-4839-4085-81f5-2b2e7cc59038 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.673281] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Waiting for the task: (returnval){ [ 905.673281] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52fca1b9-496f-deaf-908c-fe4e7e4b8638" [ 905.673281] env[61906]: _type = "Task" [ 905.673281] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.684497] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52fca1b9-496f-deaf-908c-fe4e7e4b8638, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.733935] env[61906]: DEBUG oslo_vmware.api [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333318, 'name': Rename_Task, 'duration_secs': 0.193656} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.733935] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 905.733935] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-04717625-e0a3-44f4-bf98-7f14b710918a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.742899] env[61906]: DEBUG oslo_vmware.api [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 905.742899] env[61906]: value = "task-1333319" [ 905.742899] env[61906]: _type = "Task" [ 905.742899] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.753896] env[61906]: DEBUG nova.network.neutron [-] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.755428] env[61906]: DEBUG oslo_vmware.api [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333319, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.781957] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f66c0ca3-e56f-4880-9603-e744f174d80d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.793969] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd113881-e889-4274-95ef-dd0c3cc1e697 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.834581] env[61906]: DEBUG nova.compute.manager [req-8e1bef8a-d13b-461e-b82a-8abb2ca57911 req-8a097445-5083-4471-94bc-a11f3d8f4f2f service nova] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Detach interface failed, port_id=8bcec3ac-5104-4e72-af50-08fae9dc156c, reason: Instance 75b9738e-e9b6-435a-88bb-851982828d36 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 905.980093] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3aa5781f-8be7-444b-b705-cae2fe761046 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "b1159533-c970-49d6-af42-b954b20d92fb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.588s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.043057] env[61906]: DEBUG oslo_concurrency.lockutils [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.865s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.045719] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.233s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.045990] env[61906]: DEBUG nova.objects.instance [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lazy-loading 'resources' on Instance uuid cb34a0ce-d274-4167-8367-e19809cc9e4d {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 906.070806] env[61906]: INFO nova.scheduler.client.report [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Deleted allocations for instance 9f1253e6-4727-4dca-bad2-04f0c5424cdb [ 906.184415] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52fca1b9-496f-deaf-908c-fe4e7e4b8638, 'name': SearchDatastore_Task, 'duration_secs': 0.014178} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.191631] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e807192-bf45-4619-b83d-bbbfa804369c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.202028] env[61906]: DEBUG nova.compute.manager [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 906.204737] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Waiting for the task: (returnval){ [ 906.204737] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5285ea31-5105-4e64-3d40-1b0a504ca8da" [ 906.204737] env[61906]: _type = "Task" [ 906.204737] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.219801] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5285ea31-5105-4e64-3d40-1b0a504ca8da, 'name': SearchDatastore_Task, 'duration_secs': 0.015638} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.220768] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.221613] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 0e86bfdf-9002-4a2d-a2ac-af35921b020e/0e86bfdf-9002-4a2d-a2ac-af35921b020e.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 906.221954] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fb1ac37b-aeb4-4700-9fdd-81bd34521dcd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.232534] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Waiting for the task: (returnval){ [ 906.232534] env[61906]: value = "task-1333320" [ 906.232534] env[61906]: _type = "Task" [ 906.232534] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.247905] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Task: {'id': task-1333320, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.259670] env[61906]: INFO nova.compute.manager [-] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Took 1.48 seconds to deallocate network for instance. [ 906.273556] env[61906]: DEBUG oslo_vmware.api [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333319, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.532748] env[61906]: DEBUG nova.network.neutron [req-9c3edf2d-2ba6-4430-9646-320ff0785d23 req-e8e57ee1-8a92-4514-94c2-a8e95da13355 service nova] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Updated VIF entry in instance network info cache for port 7e726302-5576-4a98-b2a5-23c4ed0be21e. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 906.533528] env[61906]: DEBUG nova.network.neutron [req-9c3edf2d-2ba6-4430-9646-320ff0785d23 req-e8e57ee1-8a92-4514-94c2-a8e95da13355 service nova] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Updating instance_info_cache with network_info: [{"id": "7e726302-5576-4a98-b2a5-23c4ed0be21e", "address": "fa:16:3e:e0:f7:dd", "network": {"id": "2240e54b-fc37-4d48-9ceb-a386504ab2c2", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-771370996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "357662e4b2e7430b82a521ed0f4392ad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5bc2dd1b-c901-4f3d-9c3a-af171c275c4b", "external-id": "nsx-vlan-transportzone-320", "segmentation_id": 320, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e726302-55", "ovs_interfaceid": "7e726302-5576-4a98-b2a5-23c4ed0be21e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.585810] env[61906]: DEBUG oslo_concurrency.lockutils [None req-26f5dfb8-be91-46a6-bb94-208c6f76987b tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "9f1253e6-4727-4dca-bad2-04f0c5424cdb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.188s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.747154] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Task: {'id': task-1333320, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.767482] env[61906]: DEBUG oslo_vmware.api [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333319, 'name': PowerOnVM_Task, 'duration_secs': 0.586335} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.768966] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 906.769818] env[61906]: INFO nova.compute.manager [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Took 8.09 seconds to spawn the instance on the hypervisor. [ 906.770051] env[61906]: DEBUG nova.compute.manager [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 906.771017] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5bf3a1c-ec1c-40e8-aea9-60bedea942a4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.775325] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8962bfb-8b39-4289-97d4-cba94243fd4d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.780617] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.789815] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68d74b13-9f05-4581-8a09-16f73606bd32 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.837575] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dd25345-4824-4e51-bd20-93ce0aa3d6c5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.848286] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-657368bb-d22e-48b6-b072-df89ca84c5c5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.868799] env[61906]: DEBUG nova.compute.provider_tree [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 906.910131] env[61906]: DEBUG nova.network.neutron [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Successfully updated port: b54c27f8-7727-4166-af82-5b6fe0d3519e {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 907.037425] env[61906]: DEBUG oslo_concurrency.lockutils [req-9c3edf2d-2ba6-4430-9646-320ff0785d23 req-e8e57ee1-8a92-4514-94c2-a8e95da13355 service nova] Releasing lock "refresh_cache-0e86bfdf-9002-4a2d-a2ac-af35921b020e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.246242] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Task: {'id': task-1333320, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.67335} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.246841] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 0e86bfdf-9002-4a2d-a2ac-af35921b020e/0e86bfdf-9002-4a2d-a2ac-af35921b020e.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 907.246950] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 907.247204] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a83ab4d7-5eee-411b-9d78-18ce28b1b4d6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.257422] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Waiting for the task: (returnval){ [ 907.257422] env[61906]: value = "task-1333321" [ 907.257422] env[61906]: _type = "Task" [ 907.257422] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.267346] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Task: {'id': task-1333321, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.308356] env[61906]: INFO nova.compute.manager [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Took 27.99 seconds to build instance. [ 907.373039] env[61906]: DEBUG nova.scheduler.client.report [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 907.414891] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "refresh_cache-e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.415098] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired lock "refresh_cache-e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.415203] env[61906]: DEBUG nova.network.neutron [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 907.769349] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Task: {'id': task-1333321, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.104076} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.769689] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 907.770712] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b2ad270-be9f-4bcf-aed5-649db4c4bc24 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.797201] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] 0e86bfdf-9002-4a2d-a2ac-af35921b020e/0e86bfdf-9002-4a2d-a2ac-af35921b020e.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 907.797654] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2f99eb8-9072-4b48-8193-d8d9c1f5effb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.814999] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7477e550-c746-40fc-b7fe-a7c212b91bad tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "1bc9f3d4-b539-4841-b5b6-eaefa0d06a72" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.152s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.823179] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Waiting for the task: (returnval){ [ 907.823179] env[61906]: value = "task-1333322" [ 907.823179] env[61906]: _type = "Task" [ 907.823179] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.834743] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Task: {'id': task-1333322, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.878232] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.832s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.881039] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.100s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.881312] env[61906]: DEBUG nova.objects.instance [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Lazy-loading 'resources' on Instance uuid 75b9738e-e9b6-435a-88bb-851982828d36 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.905073] env[61906]: INFO nova.scheduler.client.report [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Deleted allocations for instance cb34a0ce-d274-4167-8367-e19809cc9e4d [ 907.971724] env[61906]: DEBUG nova.network.neutron [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 908.153787] env[61906]: DEBUG nova.network.neutron [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Updating instance_info_cache with network_info: [{"id": "b54c27f8-7727-4166-af82-5b6fe0d3519e", "address": "fa:16:3e:ae:b6:e5", "network": {"id": "af039bdd-ffd2-483d-9655-e7df48936290", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-111531540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f7fc2042e6549cabe98838990f35842", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb54c27f8-77", "ovs_interfaceid": "b54c27f8-7727-4166-af82-5b6fe0d3519e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.335087] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Task: {'id': task-1333322, 'name': ReconfigVM_Task, 'duration_secs': 0.399317} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.335087] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Reconfigured VM instance instance-0000004a to attach disk [datastore2] 0e86bfdf-9002-4a2d-a2ac-af35921b020e/0e86bfdf-9002-4a2d-a2ac-af35921b020e.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 908.335741] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c99416d7-eee9-4a9b-84ba-9a4403e88c9f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.345105] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Waiting for the task: (returnval){ [ 908.345105] env[61906]: value = "task-1333323" [ 908.345105] env[61906]: _type = "Task" [ 908.345105] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.355609] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Task: {'id': task-1333323, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.415246] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bb2aa737-823e-4757-b6db-097385e1cdcb tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "cb34a0ce-d274-4167-8367-e19809cc9e4d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.648s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.531121] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9024d3ab-1ec0-4ee4-9d7a-4638dfff842b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.540466] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ccf9786-4055-4145-bf48-82202810a6ab {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.575891] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a598784a-281b-4223-b647-9bea5d237e8b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.584689] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af45a9ae-d6fa-4b6b-a9c5-f63b4962afac {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.600321] env[61906]: DEBUG nova.compute.provider_tree [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 908.655858] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Releasing lock "refresh_cache-e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.656211] env[61906]: DEBUG nova.compute.manager [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Instance network_info: |[{"id": "b54c27f8-7727-4166-af82-5b6fe0d3519e", "address": "fa:16:3e:ae:b6:e5", "network": {"id": "af039bdd-ffd2-483d-9655-e7df48936290", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-111531540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f7fc2042e6549cabe98838990f35842", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb54c27f8-77", "ovs_interfaceid": "b54c27f8-7727-4166-af82-5b6fe0d3519e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 908.856662] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Task: {'id': task-1333323, 'name': Rename_Task, 'duration_secs': 0.207403} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.857039] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 908.857253] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-08b7678d-9711-4f5f-bb64-d3ab4818a466 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.865533] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Waiting for the task: (returnval){ [ 908.865533] env[61906]: value = "task-1333324" [ 908.865533] env[61906]: _type = "Task" [ 908.865533] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.876256] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Task: {'id': task-1333324, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.109927] env[61906]: DEBUG nova.scheduler.client.report [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 909.377566] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Task: {'id': task-1333324, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.618056] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.737s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.642556] env[61906]: INFO nova.scheduler.client.report [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Deleted allocations for instance 75b9738e-e9b6-435a-88bb-851982828d36 [ 909.808381] env[61906]: DEBUG nova.virt.hardware [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 909.808381] env[61906]: DEBUG nova.virt.hardware [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 909.808381] env[61906]: DEBUG nova.virt.hardware [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 909.812270] env[61906]: DEBUG nova.virt.hardware [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 909.812543] env[61906]: DEBUG nova.virt.hardware [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 909.812718] env[61906]: DEBUG nova.virt.hardware [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 909.815239] env[61906]: DEBUG nova.virt.hardware [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 909.815239] env[61906]: DEBUG nova.virt.hardware [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 909.815239] env[61906]: DEBUG nova.virt.hardware [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 909.815239] env[61906]: DEBUG nova.virt.hardware [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 909.815239] env[61906]: DEBUG nova.virt.hardware [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 909.816074] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a320ef9-0f85-49d3-9b75-4cbbe9d148a1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.832682] env[61906]: DEBUG oslo_vmware.rw_handles [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524b4823-46bc-bfd8-bb2d-bc2ecac8f1ee/disk-0.vmdk. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 909.834049] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fadd5bcb-4451-4742-9131-d1cbcc64b951 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.838563] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-444ba2c1-aeb8-486b-834e-35ada3c0e29d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.854534] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:b6:e5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '19440099-773e-4a31-b82e-84a4daa5d8fe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b54c27f8-7727-4166-af82-5b6fe0d3519e', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 909.861199] env[61906]: DEBUG oslo.service.loopingcall [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 909.864155] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 909.865414] env[61906]: DEBUG nova.compute.manager [req-d07ecc42-dffb-45e3-98c8-5f6d85f7c669 req-74d8a754-5335-4d9d-83a7-e05c3b758378 service nova] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Received event network-changed-ba5a12df-7917-4b2d-ac9e-7ef1380dabf7 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 909.865606] env[61906]: DEBUG nova.compute.manager [req-d07ecc42-dffb-45e3-98c8-5f6d85f7c669 req-74d8a754-5335-4d9d-83a7-e05c3b758378 service nova] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Refreshing instance network info cache due to event network-changed-ba5a12df-7917-4b2d-ac9e-7ef1380dabf7. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 909.865822] env[61906]: DEBUG oslo_concurrency.lockutils [req-d07ecc42-dffb-45e3-98c8-5f6d85f7c669 req-74d8a754-5335-4d9d-83a7-e05c3b758378 service nova] Acquiring lock "refresh_cache-b1159533-c970-49d6-af42-b954b20d92fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.865969] env[61906]: DEBUG oslo_concurrency.lockutils [req-d07ecc42-dffb-45e3-98c8-5f6d85f7c669 req-74d8a754-5335-4d9d-83a7-e05c3b758378 service nova] Acquired lock "refresh_cache-b1159533-c970-49d6-af42-b954b20d92fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.866152] env[61906]: DEBUG nova.network.neutron [req-d07ecc42-dffb-45e3-98c8-5f6d85f7c669 req-74d8a754-5335-4d9d-83a7-e05c3b758378 service nova] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Refreshing network info cache for port ba5a12df-7917-4b2d-ac9e-7ef1380dabf7 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 909.867505] env[61906]: DEBUG oslo_vmware.rw_handles [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524b4823-46bc-bfd8-bb2d-bc2ecac8f1ee/disk-0.vmdk is in state: ready. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 909.867666] env[61906]: ERROR oslo_vmware.rw_handles [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524b4823-46bc-bfd8-bb2d-bc2ecac8f1ee/disk-0.vmdk due to incomplete transfer. [ 909.868105] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-445c4406-96d2-4b7a-ad02-98f41c556aa3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.885034] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-fc125164-a451-4e22-94b6-8253ef84255d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.898266] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Task: {'id': task-1333324, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.899599] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 909.899599] env[61906]: value = "task-1333325" [ 909.899599] env[61906]: _type = "Task" [ 909.899599] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.905557] env[61906]: DEBUG oslo_vmware.rw_handles [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524b4823-46bc-bfd8-bb2d-bc2ecac8f1ee/disk-0.vmdk. {{(pid=61906) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 909.905797] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Uploaded image 22249871-324b-4f3e-9a02-9f2cc7665edb to the Glance image server {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 909.908434] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Destroying the VM {{(pid=61906) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 909.911643] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a070d23a-bef1-4bee-b785-eacc42598cb1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.913652] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333325, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.920543] env[61906]: DEBUG oslo_vmware.api [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 909.920543] env[61906]: value = "task-1333326" [ 909.920543] env[61906]: _type = "Task" [ 909.920543] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.935025] env[61906]: DEBUG oslo_vmware.api [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333326, 'name': Destroy_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.151313] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e43d559a-8fec-44a9-8ca6-0ac8c2a9c849 tempest-ServerRescueTestJSONUnderV235-696426587 tempest-ServerRescueTestJSONUnderV235-696426587-project-member] Lock "75b9738e-e9b6-435a-88bb-851982828d36" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.519s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.402589] env[61906]: DEBUG oslo_vmware.api [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Task: {'id': task-1333324, 'name': PowerOnVM_Task, 'duration_secs': 1.352735} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.406687] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 910.406851] env[61906]: INFO nova.compute.manager [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Took 9.37 seconds to spawn the instance on the hypervisor. [ 910.407066] env[61906]: DEBUG nova.compute.manager [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 910.408154] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a82987-1327-4e20-b9d2-6e49ff02dede {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.418259] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333325, 'name': CreateVM_Task, 'duration_secs': 0.508459} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.419749] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 910.426025] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.426025] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.426025] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 910.426025] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c877c7ff-58e8-447f-9663-a12b7fc29aab {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.433841] env[61906]: DEBUG oslo_vmware.api [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 910.433841] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52b232fa-a8d0-5d0c-cde6-fcfe4caffacc" [ 910.433841] env[61906]: _type = "Task" [ 910.433841] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.437039] env[61906]: DEBUG oslo_vmware.api [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333326, 'name': Destroy_Task, 'duration_secs': 0.477371} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.440335] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Destroyed the VM [ 910.440632] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Deleting Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 910.440868] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-3f07b558-0e81-4ec9-8602-60fe74b166f4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.449224] env[61906]: DEBUG oslo_vmware.api [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52b232fa-a8d0-5d0c-cde6-fcfe4caffacc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.450857] env[61906]: DEBUG oslo_vmware.api [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 910.450857] env[61906]: value = "task-1333327" [ 910.450857] env[61906]: _type = "Task" [ 910.450857] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.461011] env[61906]: DEBUG oslo_vmware.api [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333327, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.792141] env[61906]: DEBUG nova.network.neutron [req-d07ecc42-dffb-45e3-98c8-5f6d85f7c669 req-74d8a754-5335-4d9d-83a7-e05c3b758378 service nova] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Updated VIF entry in instance network info cache for port ba5a12df-7917-4b2d-ac9e-7ef1380dabf7. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 910.792663] env[61906]: DEBUG nova.network.neutron [req-d07ecc42-dffb-45e3-98c8-5f6d85f7c669 req-74d8a754-5335-4d9d-83a7-e05c3b758378 service nova] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Updating instance_info_cache with network_info: [{"id": "ba5a12df-7917-4b2d-ac9e-7ef1380dabf7", "address": "fa:16:3e:d3:84:8a", "network": {"id": "1c55d0cc-4c30-4b89-8736-3e96c5c9c507", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-245940583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fbdcd2dc0d04e83b9f82a19bdfd0023", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bcd9d2d-25c8-41ad-9a4a-93b9029ba993", "external-id": "nsx-vlan-transportzone-734", "segmentation_id": 734, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba5a12df-79", "ovs_interfaceid": "ba5a12df-7917-4b2d-ac9e-7ef1380dabf7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.884817] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c6275799-a26c-4463-9fa5-db86f4b0c127 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "1bc9f3d4-b539-4841-b5b6-eaefa0d06a72" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.885185] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c6275799-a26c-4463-9fa5-db86f4b0c127 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "1bc9f3d4-b539-4841-b5b6-eaefa0d06a72" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.885519] env[61906]: DEBUG nova.compute.manager [None req-c6275799-a26c-4463-9fa5-db86f4b0c127 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 910.886507] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-565c2660-1dac-47ea-970b-d4e8276ea469 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.894582] env[61906]: DEBUG nova.compute.manager [None req-c6275799-a26c-4463-9fa5-db86f4b0c127 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61906) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 910.895218] env[61906]: DEBUG nova.objects.instance [None req-c6275799-a26c-4463-9fa5-db86f4b0c127 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lazy-loading 'flavor' on Instance uuid 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 910.933804] env[61906]: INFO nova.compute.manager [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Took 31.02 seconds to build instance. [ 910.950626] env[61906]: DEBUG oslo_vmware.api [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52b232fa-a8d0-5d0c-cde6-fcfe4caffacc, 'name': SearchDatastore_Task, 'duration_secs': 0.020593} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.951930] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.952286] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 910.952606] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.952867] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.953130] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 910.953636] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-65445d29-2874-4f5f-8026-ea75d1b176b6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.969407] env[61906]: DEBUG oslo_vmware.api [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333327, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.971288] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 910.971476] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 910.972458] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a04b7d35-1402-4a96-8113-3fd8dff1892b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.978763] env[61906]: DEBUG oslo_vmware.api [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 910.978763] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]523e2836-1206-1779-00bc-034899133434" [ 910.978763] env[61906]: _type = "Task" [ 910.978763] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.990103] env[61906]: DEBUG oslo_vmware.api [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]523e2836-1206-1779-00bc-034899133434, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.136801] env[61906]: DEBUG oslo_concurrency.lockutils [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "cc8870d0-eafb-4f73-aa79-b98f51370237" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.137234] env[61906]: DEBUG oslo_concurrency.lockutils [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "cc8870d0-eafb-4f73-aa79-b98f51370237" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.137490] env[61906]: DEBUG oslo_concurrency.lockutils [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "cc8870d0-eafb-4f73-aa79-b98f51370237-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.137703] env[61906]: DEBUG oslo_concurrency.lockutils [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "cc8870d0-eafb-4f73-aa79-b98f51370237-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.137885] env[61906]: DEBUG oslo_concurrency.lockutils [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "cc8870d0-eafb-4f73-aa79-b98f51370237-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.140895] env[61906]: INFO nova.compute.manager [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Terminating instance [ 911.143517] env[61906]: DEBUG nova.compute.manager [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 911.143774] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 911.144913] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36e40ab9-645b-48e1-92a8-ce055e1f0779 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.153789] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 911.154111] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9bda3d78-711b-49f6-aca3-321d9142bff9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.279643] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 911.280026] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 911.280190] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Deleting the datastore file [datastore1] cc8870d0-eafb-4f73-aa79-b98f51370237 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 911.280364] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0fc65877-e0ca-409e-be07-2663bdbf849e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.288857] env[61906]: DEBUG oslo_vmware.api [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 911.288857] env[61906]: value = "task-1333329" [ 911.288857] env[61906]: _type = "Task" [ 911.288857] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.298373] env[61906]: DEBUG oslo_concurrency.lockutils [req-d07ecc42-dffb-45e3-98c8-5f6d85f7c669 req-74d8a754-5335-4d9d-83a7-e05c3b758378 service nova] Releasing lock "refresh_cache-b1159533-c970-49d6-af42-b954b20d92fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.306471] env[61906]: DEBUG oslo_vmware.api [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333329, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.403637] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6275799-a26c-4463-9fa5-db86f4b0c127 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 911.404146] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9b0bde0c-34a0-4ede-8d96-0e539e11803a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.413281] env[61906]: DEBUG oslo_vmware.api [None req-c6275799-a26c-4463-9fa5-db86f4b0c127 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 911.413281] env[61906]: value = "task-1333330" [ 911.413281] env[61906]: _type = "Task" [ 911.413281] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.428894] env[61906]: DEBUG oslo_vmware.api [None req-c6275799-a26c-4463-9fa5-db86f4b0c127 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333330, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.438632] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1c1438a3-a921-4a20-8368-2795b97756fd tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Lock "0e86bfdf-9002-4a2d-a2ac-af35921b020e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.004s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.464117] env[61906]: DEBUG oslo_vmware.api [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333327, 'name': RemoveSnapshot_Task, 'duration_secs': 0.774822} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.464117] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Deleted Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 911.464385] env[61906]: INFO nova.compute.manager [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Took 14.03 seconds to snapshot the instance on the hypervisor. [ 911.491674] env[61906]: DEBUG oslo_vmware.api [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]523e2836-1206-1779-00bc-034899133434, 'name': SearchDatastore_Task, 'duration_secs': 0.018797} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.492661] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb249003-1c16-4391-b3a2-69148f3a07dd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.500359] env[61906]: DEBUG oslo_vmware.api [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 911.500359] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]529bd560-1626-55e5-0596-9dc7c617f834" [ 911.500359] env[61906]: _type = "Task" [ 911.500359] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.511554] env[61906]: DEBUG oslo_vmware.api [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]529bd560-1626-55e5-0596-9dc7c617f834, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.618862] env[61906]: DEBUG oslo_concurrency.lockutils [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "0488c6eb-b510-4d33-ab9f-ebfdab63fb3a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.619367] env[61906]: DEBUG oslo_concurrency.lockutils [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "0488c6eb-b510-4d33-ab9f-ebfdab63fb3a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.806755] env[61906]: DEBUG oslo_vmware.api [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333329, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.318884} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.807168] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 911.807397] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 911.807555] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 911.807742] env[61906]: INFO nova.compute.manager [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Took 0.66 seconds to destroy the instance on the hypervisor. [ 911.807999] env[61906]: DEBUG oslo.service.loopingcall [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 911.808227] env[61906]: DEBUG nova.compute.manager [-] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 911.808356] env[61906]: DEBUG nova.network.neutron [-] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 911.926284] env[61906]: DEBUG oslo_vmware.api [None req-c6275799-a26c-4463-9fa5-db86f4b0c127 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333330, 'name': PowerOffVM_Task, 'duration_secs': 0.219803} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.929707] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6275799-a26c-4463-9fa5-db86f4b0c127 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 911.929707] env[61906]: DEBUG nova.compute.manager [None req-c6275799-a26c-4463-9fa5-db86f4b0c127 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 911.930860] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-327f5c30-7321-4796-bf76-f8a5d8402dfd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.935161] env[61906]: DEBUG oslo_concurrency.lockutils [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "ad2db007-e7aa-4c66-9020-ac4acfcff25a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.935417] env[61906]: DEBUG oslo_concurrency.lockutils [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "ad2db007-e7aa-4c66-9020-ac4acfcff25a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.971384] env[61906]: DEBUG nova.compute.manager [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Instance disappeared during snapshot {{(pid=61906) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 911.982190] env[61906]: DEBUG nova.compute.manager [req-0033adce-5ae0-4000-b0d8-ba9ee80e4336 req-21922120-3955-4ad2-9fe9-b76621b559ef service nova] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Received event network-vif-plugged-b54c27f8-7727-4166-af82-5b6fe0d3519e {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.982833] env[61906]: DEBUG oslo_concurrency.lockutils [req-0033adce-5ae0-4000-b0d8-ba9ee80e4336 req-21922120-3955-4ad2-9fe9-b76621b559ef service nova] Acquiring lock "e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.983456] env[61906]: DEBUG oslo_concurrency.lockutils [req-0033adce-5ae0-4000-b0d8-ba9ee80e4336 req-21922120-3955-4ad2-9fe9-b76621b559ef service nova] Lock "e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.983456] env[61906]: DEBUG oslo_concurrency.lockutils [req-0033adce-5ae0-4000-b0d8-ba9ee80e4336 req-21922120-3955-4ad2-9fe9-b76621b559ef service nova] Lock "e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.985011] env[61906]: DEBUG nova.compute.manager [req-0033adce-5ae0-4000-b0d8-ba9ee80e4336 req-21922120-3955-4ad2-9fe9-b76621b559ef service nova] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] No waiting events found dispatching network-vif-plugged-b54c27f8-7727-4166-af82-5b6fe0d3519e {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 911.985011] env[61906]: WARNING nova.compute.manager [req-0033adce-5ae0-4000-b0d8-ba9ee80e4336 req-21922120-3955-4ad2-9fe9-b76621b559ef service nova] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Received unexpected event network-vif-plugged-b54c27f8-7727-4166-af82-5b6fe0d3519e for instance with vm_state building and task_state spawning. [ 911.985011] env[61906]: DEBUG nova.compute.manager [req-0033adce-5ae0-4000-b0d8-ba9ee80e4336 req-21922120-3955-4ad2-9fe9-b76621b559ef service nova] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Received event network-changed-b54c27f8-7727-4166-af82-5b6fe0d3519e {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.985011] env[61906]: DEBUG nova.compute.manager [req-0033adce-5ae0-4000-b0d8-ba9ee80e4336 req-21922120-3955-4ad2-9fe9-b76621b559ef service nova] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Refreshing instance network info cache due to event network-changed-b54c27f8-7727-4166-af82-5b6fe0d3519e. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 911.985011] env[61906]: DEBUG oslo_concurrency.lockutils [req-0033adce-5ae0-4000-b0d8-ba9ee80e4336 req-21922120-3955-4ad2-9fe9-b76621b559ef service nova] Acquiring lock "refresh_cache-e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.985011] env[61906]: DEBUG oslo_concurrency.lockutils [req-0033adce-5ae0-4000-b0d8-ba9ee80e4336 req-21922120-3955-4ad2-9fe9-b76621b559ef service nova] Acquired lock "refresh_cache-e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.985011] env[61906]: DEBUG nova.network.neutron [req-0033adce-5ae0-4000-b0d8-ba9ee80e4336 req-21922120-3955-4ad2-9fe9-b76621b559ef service nova] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Refreshing network info cache for port b54c27f8-7727-4166-af82-5b6fe0d3519e {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 911.989349] env[61906]: DEBUG nova.compute.manager [None req-92ef1d96-3c0d-47db-be24-7f25504d1e24 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Image not found during clean up 22249871-324b-4f3e-9a02-9f2cc7665edb {{(pid=61906) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4500}} [ 912.013396] env[61906]: DEBUG oslo_vmware.api [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]529bd560-1626-55e5-0596-9dc7c617f834, 'name': SearchDatastore_Task, 'duration_secs': 0.013034} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.013977] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.017286] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15/e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 912.017623] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c2f25adb-9965-4529-add6-15c685ca226f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.042552] env[61906]: DEBUG oslo_vmware.api [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 912.042552] env[61906]: value = "task-1333331" [ 912.042552] env[61906]: _type = "Task" [ 912.042552] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.056421] env[61906]: DEBUG oslo_vmware.api [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333331, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.123634] env[61906]: DEBUG nova.compute.manager [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 912.438491] env[61906]: DEBUG nova.compute.manager [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 912.454406] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c6275799-a26c-4463-9fa5-db86f4b0c127 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "1bc9f3d4-b539-4841-b5b6-eaefa0d06a72" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.569s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.556268] env[61906]: DEBUG oslo_vmware.api [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333331, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.623783] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "1a335019-915b-4065-b22f-d2f0af728d6f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.624072] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "1a335019-915b-4065-b22f-d2f0af728d6f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.652238] env[61906]: DEBUG oslo_concurrency.lockutils [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.652644] env[61906]: DEBUG oslo_concurrency.lockutils [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.655037] env[61906]: INFO nova.compute.claims [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 912.657885] env[61906]: DEBUG nova.network.neutron [-] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.723909] env[61906]: DEBUG oslo_concurrency.lockutils [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Acquiring lock "0e86bfdf-9002-4a2d-a2ac-af35921b020e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.724104] env[61906]: DEBUG oslo_concurrency.lockutils [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Lock "0e86bfdf-9002-4a2d-a2ac-af35921b020e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.724333] env[61906]: DEBUG oslo_concurrency.lockutils [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Acquiring lock "0e86bfdf-9002-4a2d-a2ac-af35921b020e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.724528] env[61906]: DEBUG oslo_concurrency.lockutils [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Lock "0e86bfdf-9002-4a2d-a2ac-af35921b020e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.724724] env[61906]: DEBUG oslo_concurrency.lockutils [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Lock "0e86bfdf-9002-4a2d-a2ac-af35921b020e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.727196] env[61906]: INFO nova.compute.manager [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Terminating instance [ 912.730286] env[61906]: DEBUG nova.compute.manager [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 912.730365] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 912.731885] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa9c2898-ad14-405a-82ce-c7a85eae38e6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.740977] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 912.741281] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3df26e97-000e-49c4-ae72-cf6872fedb34 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.750506] env[61906]: DEBUG oslo_vmware.api [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Waiting for the task: (returnval){ [ 912.750506] env[61906]: value = "task-1333332" [ 912.750506] env[61906]: _type = "Task" [ 912.750506] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.762034] env[61906]: DEBUG oslo_vmware.api [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Task: {'id': task-1333332, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.779243] env[61906]: DEBUG nova.network.neutron [req-0033adce-5ae0-4000-b0d8-ba9ee80e4336 req-21922120-3955-4ad2-9fe9-b76621b559ef service nova] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Updated VIF entry in instance network info cache for port b54c27f8-7727-4166-af82-5b6fe0d3519e. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 912.779640] env[61906]: DEBUG nova.network.neutron [req-0033adce-5ae0-4000-b0d8-ba9ee80e4336 req-21922120-3955-4ad2-9fe9-b76621b559ef service nova] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Updating instance_info_cache with network_info: [{"id": "b54c27f8-7727-4166-af82-5b6fe0d3519e", "address": "fa:16:3e:ae:b6:e5", "network": {"id": "af039bdd-ffd2-483d-9655-e7df48936290", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-111531540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f7fc2042e6549cabe98838990f35842", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb54c27f8-77", "ovs_interfaceid": "b54c27f8-7727-4166-af82-5b6fe0d3519e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.962964] env[61906]: DEBUG oslo_concurrency.lockutils [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.056728] env[61906]: DEBUG oslo_vmware.api [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333331, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.727463} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.057109] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15/e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 913.057416] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 913.058867] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-de8dce00-6980-4d51-9c23-ab026c70088d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.070508] env[61906]: DEBUG oslo_vmware.api [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 913.070508] env[61906]: value = "task-1333333" [ 913.070508] env[61906]: _type = "Task" [ 913.070508] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.081533] env[61906]: DEBUG oslo_vmware.api [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333333, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.136162] env[61906]: DEBUG nova.compute.manager [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 913.162402] env[61906]: INFO nova.compute.manager [-] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Took 1.35 seconds to deallocate network for instance. [ 913.262582] env[61906]: DEBUG oslo_vmware.api [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Task: {'id': task-1333332, 'name': PowerOffVM_Task, 'duration_secs': 0.206506} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.262827] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 913.263010] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 913.263338] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c4ca7e12-4aff-45a9-9083-c972939ab674 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.282508] env[61906]: DEBUG oslo_concurrency.lockutils [req-0033adce-5ae0-4000-b0d8-ba9ee80e4336 req-21922120-3955-4ad2-9fe9-b76621b559ef service nova] Releasing lock "refresh_cache-e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.335283] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 913.335448] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 913.335637] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Deleting the datastore file [datastore2] 0e86bfdf-9002-4a2d-a2ac-af35921b020e {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 913.335945] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8c015b06-9552-48bb-87a0-368ecf229fd8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.342915] env[61906]: DEBUG oslo_vmware.api [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Waiting for the task: (returnval){ [ 913.342915] env[61906]: value = "task-1333335" [ 913.342915] env[61906]: _type = "Task" [ 913.342915] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.351883] env[61906]: DEBUG oslo_vmware.api [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Task: {'id': task-1333335, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.565293] env[61906]: DEBUG oslo_concurrency.lockutils [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "1bc9f3d4-b539-4841-b5b6-eaefa0d06a72" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.565766] env[61906]: DEBUG oslo_concurrency.lockutils [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "1bc9f3d4-b539-4841-b5b6-eaefa0d06a72" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.565766] env[61906]: DEBUG oslo_concurrency.lockutils [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "1bc9f3d4-b539-4841-b5b6-eaefa0d06a72-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.565915] env[61906]: DEBUG oslo_concurrency.lockutils [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "1bc9f3d4-b539-4841-b5b6-eaefa0d06a72-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.566016] env[61906]: DEBUG oslo_concurrency.lockutils [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "1bc9f3d4-b539-4841-b5b6-eaefa0d06a72-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.568475] env[61906]: INFO nova.compute.manager [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Terminating instance [ 913.570435] env[61906]: DEBUG nova.compute.manager [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 913.571056] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 913.571920] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-499f9530-ffda-4de7-afb1-f11289e18b38 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.585868] env[61906]: DEBUG oslo_vmware.api [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333333, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088596} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.588221] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 913.588472] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 913.589215] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e0e9846-2361-4ad5-8133-6461726dd987 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.591759] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fb91af84-5ab1-4bb3-aa9d-05dafadfe85c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.619321] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15/e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 913.619683] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2230f1a5-1e71-4612-8028-2750bfe9762b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.645459] env[61906]: DEBUG oslo_vmware.api [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 913.645459] env[61906]: value = "task-1333337" [ 913.645459] env[61906]: _type = "Task" [ 913.645459] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.655033] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.668267] env[61906]: DEBUG oslo_vmware.api [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333337, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.671463] env[61906]: DEBUG oslo_concurrency.lockutils [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.686076] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 913.686317] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 913.686550] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Deleting the datastore file [datastore2] 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 913.686760] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-602d2222-659e-41af-84d1-68e4b873249e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.694840] env[61906]: DEBUG oslo_vmware.api [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 913.694840] env[61906]: value = "task-1333338" [ 913.694840] env[61906]: _type = "Task" [ 913.694840] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.706792] env[61906]: DEBUG oslo_vmware.api [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333338, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.840240] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1984d58e-6ce2-4006-908d-c72ecfbe1e06 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.849628] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a4f051-d777-4221-b674-e0b358c51328 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.855953] env[61906]: DEBUG oslo_vmware.api [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Task: {'id': task-1333335, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.24728} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.856680] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 913.856883] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 913.857125] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 913.857343] env[61906]: INFO nova.compute.manager [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Took 1.13 seconds to destroy the instance on the hypervisor. [ 913.857618] env[61906]: DEBUG oslo.service.loopingcall [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 913.857821] env[61906]: DEBUG nova.compute.manager [-] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 913.857920] env[61906]: DEBUG nova.network.neutron [-] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 913.885669] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0cee951-5cf0-4359-a273-1980807e94aa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.895503] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8099c9b-0573-4b1b-a46e-942be442de86 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.913355] env[61906]: DEBUG nova.compute.provider_tree [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 913.987643] env[61906]: DEBUG oslo_concurrency.lockutils [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquiring lock "11cb9d89-2304-45a6-b4bb-fcf8417a0518" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.988012] env[61906]: DEBUG oslo_concurrency.lockutils [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Lock "11cb9d89-2304-45a6-b4bb-fcf8417a0518" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.009538] env[61906]: DEBUG nova.compute.manager [req-d95736df-9181-4e98-a09d-0479912e8287 req-a34812cf-8872-4b88-93c4-bac9911951c0 service nova] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Received event network-vif-deleted-58de857e-fc72-4f45-9da6-93623ceb23bc {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 914.157011] env[61906]: DEBUG oslo_vmware.api [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333337, 'name': ReconfigVM_Task, 'duration_secs': 0.332707} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.157515] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Reconfigured VM instance instance-0000004b to attach disk [datastore1] e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15/e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 914.158210] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-30cba4e8-8e13-4849-9b01-cb0b54c22e39 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.166508] env[61906]: DEBUG oslo_vmware.api [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 914.166508] env[61906]: value = "task-1333339" [ 914.166508] env[61906]: _type = "Task" [ 914.166508] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.176766] env[61906]: DEBUG oslo_vmware.api [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333339, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.207120] env[61906]: DEBUG oslo_vmware.api [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333338, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.254171] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquiring lock "8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.254435] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Lock "8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.416849] env[61906]: DEBUG nova.scheduler.client.report [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 914.490385] env[61906]: DEBUG nova.compute.manager [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 914.639696] env[61906]: DEBUG nova.network.neutron [-] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.677382] env[61906]: DEBUG oslo_vmware.api [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333339, 'name': Rename_Task, 'duration_secs': 0.176498} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.678114] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 914.678114] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b03b384c-162f-4a06-ba99-e5da6f27937a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.685961] env[61906]: DEBUG oslo_vmware.api [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 914.685961] env[61906]: value = "task-1333340" [ 914.685961] env[61906]: _type = "Task" [ 914.685961] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.694549] env[61906]: DEBUG oslo_vmware.api [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333340, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.705468] env[61906]: DEBUG oslo_vmware.api [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333338, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.618025} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.705784] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 914.705978] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 914.706180] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 914.706360] env[61906]: INFO nova.compute.manager [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Took 1.14 seconds to destroy the instance on the hypervisor. [ 914.706615] env[61906]: DEBUG oslo.service.loopingcall [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.706808] env[61906]: DEBUG nova.compute.manager [-] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 914.706905] env[61906]: DEBUG nova.network.neutron [-] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 914.757249] env[61906]: DEBUG nova.compute.manager [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 914.922208] env[61906]: DEBUG oslo_concurrency.lockutils [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.269s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.922799] env[61906]: DEBUG nova.compute.manager [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 914.925708] env[61906]: DEBUG oslo_concurrency.lockutils [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.963s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.927218] env[61906]: INFO nova.compute.claims [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 915.013552] env[61906]: DEBUG oslo_concurrency.lockutils [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.107647] env[61906]: DEBUG oslo_vmware.rw_handles [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523d516e-d0c1-6da1-8a04-13c0867afd52/disk-0.vmdk. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 915.108613] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-459eae3a-8908-4cad-ada9-e9190f1f0d6f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.116072] env[61906]: DEBUG oslo_vmware.rw_handles [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523d516e-d0c1-6da1-8a04-13c0867afd52/disk-0.vmdk is in state: ready. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 915.116265] env[61906]: ERROR oslo_vmware.rw_handles [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523d516e-d0c1-6da1-8a04-13c0867afd52/disk-0.vmdk due to incomplete transfer. [ 915.116511] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-637cf249-3839-44f2-a39f-cc09943869c7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.124986] env[61906]: DEBUG oslo_vmware.rw_handles [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523d516e-d0c1-6da1-8a04-13c0867afd52/disk-0.vmdk. {{(pid=61906) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 915.125170] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Uploaded image 4efc4cce-9185-4c01-8fa1-4d7a2414b871 to the Glance image server {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 915.126777] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Destroying the VM {{(pid=61906) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 915.127071] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e1c0a561-e04b-46b4-9183-9b7b2ebd107b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.136040] env[61906]: DEBUG oslo_vmware.api [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Waiting for the task: (returnval){ [ 915.136040] env[61906]: value = "task-1333341" [ 915.136040] env[61906]: _type = "Task" [ 915.136040] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.144875] env[61906]: INFO nova.compute.manager [-] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Took 1.29 seconds to deallocate network for instance. [ 915.145226] env[61906]: DEBUG oslo_vmware.api [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333341, 'name': Destroy_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.197618] env[61906]: DEBUG oslo_vmware.api [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333340, 'name': PowerOnVM_Task, 'duration_secs': 0.47276} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.197876] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 915.198054] env[61906]: INFO nova.compute.manager [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Took 9.00 seconds to spawn the instance on the hypervisor. [ 915.198241] env[61906]: DEBUG nova.compute.manager [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 915.199289] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c7d718b-da3b-4951-b87e-11659e5f64d7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.276344] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.432082] env[61906]: DEBUG nova.compute.utils [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 915.437293] env[61906]: DEBUG nova.compute.manager [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 915.437429] env[61906]: DEBUG nova.network.neutron [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 915.439459] env[61906]: DEBUG nova.network.neutron [-] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.509485] env[61906]: DEBUG nova.policy [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b49e5a6aeb1d43e290a645e14861b889', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '53644a4143a24759a9ff2b5e28b84fb4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 915.645381] env[61906]: DEBUG oslo_vmware.api [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333341, 'name': Destroy_Task, 'duration_secs': 0.301131} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.645668] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Destroyed the VM [ 915.645897] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Deleting Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 915.646171] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-13e8cc10-6e85-4907-ab21-a318e98c10e1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.650869] env[61906]: DEBUG oslo_concurrency.lockutils [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.653683] env[61906]: DEBUG oslo_vmware.api [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Waiting for the task: (returnval){ [ 915.653683] env[61906]: value = "task-1333342" [ 915.653683] env[61906]: _type = "Task" [ 915.653683] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.664781] env[61906]: DEBUG oslo_vmware.api [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333342, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.720616] env[61906]: INFO nova.compute.manager [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Took 25.79 seconds to build instance. [ 915.844843] env[61906]: DEBUG nova.network.neutron [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Successfully created port: b0492e76-fbe6-486e-8fe8-539b1e22ab50 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 915.941082] env[61906]: DEBUG nova.compute.manager [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 915.944284] env[61906]: INFO nova.compute.manager [-] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Took 1.24 seconds to deallocate network for instance. [ 916.045669] env[61906]: DEBUG nova.compute.manager [req-56516652-5d20-4f74-bb71-44da18a5f29c req-517cafff-2ca9-4658-a7f1-cad5bdc99f74 service nova] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Received event network-vif-deleted-7e726302-5576-4a98-b2a5-23c4ed0be21e {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 916.045669] env[61906]: DEBUG nova.compute.manager [req-56516652-5d20-4f74-bb71-44da18a5f29c req-517cafff-2ca9-4658-a7f1-cad5bdc99f74 service nova] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Received event network-vif-deleted-1af3a42f-34db-4c40-8efc-330a284d1f2c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 916.164838] env[61906]: DEBUG oslo_vmware.api [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333342, 'name': RemoveSnapshot_Task} progress is 43%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.166422] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9baa727e-2044-47e2-ab8e-4bc3e9039e21 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.173919] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f73eec41-4b70-4da0-8d2f-c35dc612d882 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.204576] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32a0c5dc-1761-4f6d-aa3b-ba5993af9212 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.213042] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25834a9f-cdc7-4922-b96f-c67b62a49084 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.226592] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ed4dcde-3dd3-4a18-bfc1-e380ff3202c8 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.312s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.227091] env[61906]: DEBUG nova.compute.provider_tree [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.450770] env[61906]: DEBUG oslo_concurrency.lockutils [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.668239] env[61906]: DEBUG oslo_vmware.api [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333342, 'name': RemoveSnapshot_Task, 'duration_secs': 0.673065} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.668439] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Deleted Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 916.668584] env[61906]: INFO nova.compute.manager [None req-7673bc16-0670-48c2-82cd-4d5e9bc7c007 tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Took 17.04 seconds to snapshot the instance on the hypervisor. [ 916.730941] env[61906]: DEBUG nova.scheduler.client.report [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 916.951736] env[61906]: DEBUG nova.compute.manager [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 916.975079] env[61906]: DEBUG nova.virt.hardware [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 916.975339] env[61906]: DEBUG nova.virt.hardware [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 916.975507] env[61906]: DEBUG nova.virt.hardware [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 916.975698] env[61906]: DEBUG nova.virt.hardware [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 916.975850] env[61906]: DEBUG nova.virt.hardware [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 916.976008] env[61906]: DEBUG nova.virt.hardware [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 916.976383] env[61906]: DEBUG nova.virt.hardware [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 916.976561] env[61906]: DEBUG nova.virt.hardware [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 916.976736] env[61906]: DEBUG nova.virt.hardware [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 916.976906] env[61906]: DEBUG nova.virt.hardware [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 916.977103] env[61906]: DEBUG nova.virt.hardware [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 916.977957] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-264964c5-ed34-487e-b4bc-74c4d330bfe1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.988107] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-177c43c0-22b3-4395-9d86-956aa46b1427 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.236413] env[61906]: DEBUG oslo_concurrency.lockutils [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.310s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.237177] env[61906]: DEBUG nova.compute.manager [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 917.240743] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.586s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.243991] env[61906]: INFO nova.compute.claims [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 917.417756] env[61906]: DEBUG nova.network.neutron [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Successfully updated port: b0492e76-fbe6-486e-8fe8-539b1e22ab50 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 917.751752] env[61906]: DEBUG nova.compute.utils [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 917.753184] env[61906]: DEBUG nova.compute.manager [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 917.753753] env[61906]: DEBUG nova.network.neutron [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 917.779863] env[61906]: DEBUG nova.compute.manager [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Stashing vm_state: active {{(pid=61906) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 917.792370] env[61906]: DEBUG nova.policy [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc6b374325bc4ef9882e5f84ddd62fd1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f2d2198383d74dac90aa9727a62176df', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 917.920387] env[61906]: DEBUG oslo_concurrency.lockutils [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "refresh_cache-0488c6eb-b510-4d33-ab9f-ebfdab63fb3a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.920580] env[61906]: DEBUG oslo_concurrency.lockutils [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired lock "refresh_cache-0488c6eb-b510-4d33-ab9f-ebfdab63fb3a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.920719] env[61906]: DEBUG nova.network.neutron [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 918.064471] env[61906]: DEBUG nova.network.neutron [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Successfully created port: d6afef85-5b45-461f-9671-7a187b8022fc {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 918.075384] env[61906]: DEBUG nova.compute.manager [req-8b5d6d30-5a53-4372-b756-8558e6e211ec req-a7ebf362-86a4-4bf5-868c-99dca8df8c4f service nova] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Received event network-vif-plugged-b0492e76-fbe6-486e-8fe8-539b1e22ab50 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 918.075384] env[61906]: DEBUG oslo_concurrency.lockutils [req-8b5d6d30-5a53-4372-b756-8558e6e211ec req-a7ebf362-86a4-4bf5-868c-99dca8df8c4f service nova] Acquiring lock "0488c6eb-b510-4d33-ab9f-ebfdab63fb3a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.075384] env[61906]: DEBUG oslo_concurrency.lockutils [req-8b5d6d30-5a53-4372-b756-8558e6e211ec req-a7ebf362-86a4-4bf5-868c-99dca8df8c4f service nova] Lock "0488c6eb-b510-4d33-ab9f-ebfdab63fb3a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.075724] env[61906]: DEBUG oslo_concurrency.lockutils [req-8b5d6d30-5a53-4372-b756-8558e6e211ec req-a7ebf362-86a4-4bf5-868c-99dca8df8c4f service nova] Lock "0488c6eb-b510-4d33-ab9f-ebfdab63fb3a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.075724] env[61906]: DEBUG nova.compute.manager [req-8b5d6d30-5a53-4372-b756-8558e6e211ec req-a7ebf362-86a4-4bf5-868c-99dca8df8c4f service nova] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] No waiting events found dispatching network-vif-plugged-b0492e76-fbe6-486e-8fe8-539b1e22ab50 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 918.079113] env[61906]: WARNING nova.compute.manager [req-8b5d6d30-5a53-4372-b756-8558e6e211ec req-a7ebf362-86a4-4bf5-868c-99dca8df8c4f service nova] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Received unexpected event network-vif-plugged-b0492e76-fbe6-486e-8fe8-539b1e22ab50 for instance with vm_state building and task_state spawning. [ 918.079113] env[61906]: DEBUG nova.compute.manager [req-8b5d6d30-5a53-4372-b756-8558e6e211ec req-a7ebf362-86a4-4bf5-868c-99dca8df8c4f service nova] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Received event network-changed-b0492e76-fbe6-486e-8fe8-539b1e22ab50 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 918.079113] env[61906]: DEBUG nova.compute.manager [req-8b5d6d30-5a53-4372-b756-8558e6e211ec req-a7ebf362-86a4-4bf5-868c-99dca8df8c4f service nova] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Refreshing instance network info cache due to event network-changed-b0492e76-fbe6-486e-8fe8-539b1e22ab50. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 918.079113] env[61906]: DEBUG oslo_concurrency.lockutils [req-8b5d6d30-5a53-4372-b756-8558e6e211ec req-a7ebf362-86a4-4bf5-868c-99dca8df8c4f service nova] Acquiring lock "refresh_cache-0488c6eb-b510-4d33-ab9f-ebfdab63fb3a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.256946] env[61906]: DEBUG nova.compute.manager [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 918.297363] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.395926] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Acquiring lock "d60575a6-af13-4d81-95d1-081d0f3e2ce6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.396495] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Lock "d60575a6-af13-4d81-95d1-081d0f3e2ce6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.397029] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Acquiring lock "d60575a6-af13-4d81-95d1-081d0f3e2ce6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.397465] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Lock "d60575a6-af13-4d81-95d1-081d0f3e2ce6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.397818] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Lock "d60575a6-af13-4d81-95d1-081d0f3e2ce6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.400977] env[61906]: INFO nova.compute.manager [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Terminating instance [ 918.403412] env[61906]: DEBUG nova.compute.manager [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 918.403652] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 918.404663] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53590371-9e1b-4fef-8226-8e7fe86327b2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.415036] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 918.415036] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-84446788-893b-4f9a-9cc5-281d0633c5d4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.423662] env[61906]: DEBUG oslo_vmware.api [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Waiting for the task: (returnval){ [ 918.423662] env[61906]: value = "task-1333343" [ 918.423662] env[61906]: _type = "Task" [ 918.423662] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.439012] env[61906]: DEBUG oslo_vmware.api [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333343, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.461344] env[61906]: DEBUG nova.network.neutron [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 918.464893] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b7efe26-a0ad-43ae-b5ab-df10ebb7e1a1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.472619] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ec135c5-645e-4cde-b8e3-1d32e3860f5d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.504818] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78ee3fef-daf9-48ea-9206-b1b9765bc2b0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.513563] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac9ba6ce-9f69-402d-b3da-592cae0872a0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.528645] env[61906]: DEBUG nova.compute.provider_tree [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.660065] env[61906]: DEBUG nova.network.neutron [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Updating instance_info_cache with network_info: [{"id": "b0492e76-fbe6-486e-8fe8-539b1e22ab50", "address": "fa:16:3e:0f:03:18", "network": {"id": "bc9185e9-f26b-489c-84ee-e6db41ac6283", "bridge": "br-int", "label": "tempest-ServersTestJSON-1167747148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53644a4143a24759a9ff2b5e28b84fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0492e76-fb", "ovs_interfaceid": "b0492e76-fbe6-486e-8fe8-539b1e22ab50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.943193] env[61906]: DEBUG oslo_vmware.api [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333343, 'name': PowerOffVM_Task, 'duration_secs': 0.193314} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.943569] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 918.943864] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 918.944363] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c13f0dde-9e0a-46d1-8cdb-5fb6d35ca6c3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.034035] env[61906]: DEBUG nova.scheduler.client.report [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 919.047119] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 919.047477] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 919.048226] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Deleting the datastore file [datastore2] d60575a6-af13-4d81-95d1-081d0f3e2ce6 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 919.048226] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d26dbc7c-d3a5-4220-be1a-10324f0f740f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.057435] env[61906]: DEBUG oslo_vmware.api [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Waiting for the task: (returnval){ [ 919.057435] env[61906]: value = "task-1333345" [ 919.057435] env[61906]: _type = "Task" [ 919.057435] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.067127] env[61906]: DEBUG oslo_vmware.api [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333345, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.165776] env[61906]: DEBUG oslo_concurrency.lockutils [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Releasing lock "refresh_cache-0488c6eb-b510-4d33-ab9f-ebfdab63fb3a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.166129] env[61906]: DEBUG nova.compute.manager [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Instance network_info: |[{"id": "b0492e76-fbe6-486e-8fe8-539b1e22ab50", "address": "fa:16:3e:0f:03:18", "network": {"id": "bc9185e9-f26b-489c-84ee-e6db41ac6283", "bridge": "br-int", "label": "tempest-ServersTestJSON-1167747148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53644a4143a24759a9ff2b5e28b84fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0492e76-fb", "ovs_interfaceid": "b0492e76-fbe6-486e-8fe8-539b1e22ab50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 919.166466] env[61906]: DEBUG oslo_concurrency.lockutils [req-8b5d6d30-5a53-4372-b756-8558e6e211ec req-a7ebf362-86a4-4bf5-868c-99dca8df8c4f service nova] Acquired lock "refresh_cache-0488c6eb-b510-4d33-ab9f-ebfdab63fb3a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.166657] env[61906]: DEBUG nova.network.neutron [req-8b5d6d30-5a53-4372-b756-8558e6e211ec req-a7ebf362-86a4-4bf5-868c-99dca8df8c4f service nova] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Refreshing network info cache for port b0492e76-fbe6-486e-8fe8-539b1e22ab50 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 919.167852] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0f:03:18', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b0492e76-fbe6-486e-8fe8-539b1e22ab50', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 919.175323] env[61906]: DEBUG oslo.service.loopingcall [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 919.175767] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 919.175997] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d021a1d1-f22a-4e15-b8e9-d80cf738f454 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.197338] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 919.197338] env[61906]: value = "task-1333346" [ 919.197338] env[61906]: _type = "Task" [ 919.197338] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.205402] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333346, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.265922] env[61906]: DEBUG nova.compute.manager [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 919.290969] env[61906]: DEBUG nova.virt.hardware [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 919.291254] env[61906]: DEBUG nova.virt.hardware [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 919.291417] env[61906]: DEBUG nova.virt.hardware [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 919.291601] env[61906]: DEBUG nova.virt.hardware [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 919.291751] env[61906]: DEBUG nova.virt.hardware [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 919.291900] env[61906]: DEBUG nova.virt.hardware [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 919.292127] env[61906]: DEBUG nova.virt.hardware [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 919.292317] env[61906]: DEBUG nova.virt.hardware [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 919.292499] env[61906]: DEBUG nova.virt.hardware [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 919.292665] env[61906]: DEBUG nova.virt.hardware [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 919.292845] env[61906]: DEBUG nova.virt.hardware [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 919.293714] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd004a0d-f625-4698-a1e0-d0cc867a6f3b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.302414] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a83a2658-f9ae-4267-9962-25ea911df760 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.537434] env[61906]: DEBUG nova.network.neutron [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Successfully updated port: d6afef85-5b45-461f-9671-7a187b8022fc {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 919.539614] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.299s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.540019] env[61906]: DEBUG nova.compute.manager [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 919.542700] env[61906]: DEBUG oslo_concurrency.lockutils [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.871s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.543286] env[61906]: DEBUG nova.objects.instance [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lazy-loading 'resources' on Instance uuid cc8870d0-eafb-4f73-aa79-b98f51370237 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.567966] env[61906]: DEBUG oslo_vmware.api [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Task: {'id': task-1333345, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140358} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.568226] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 919.568416] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 919.568593] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 919.568766] env[61906]: INFO nova.compute.manager [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Took 1.17 seconds to destroy the instance on the hypervisor. [ 919.569009] env[61906]: DEBUG oslo.service.loopingcall [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 919.569209] env[61906]: DEBUG nova.compute.manager [-] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 919.569341] env[61906]: DEBUG nova.network.neutron [-] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 919.707548] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333346, 'name': CreateVM_Task, 'duration_secs': 0.328027} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.707722] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 919.708422] env[61906]: DEBUG oslo_concurrency.lockutils [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.708590] env[61906]: DEBUG oslo_concurrency.lockutils [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.708914] env[61906]: DEBUG oslo_concurrency.lockutils [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 919.709193] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76b09127-f747-4149-99ba-50f78b06d77e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.714175] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 919.714175] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5268acfc-6be7-2234-c2b0-5aa0ac99f8f1" [ 919.714175] env[61906]: _type = "Task" [ 919.714175] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.724578] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5268acfc-6be7-2234-c2b0-5aa0ac99f8f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.916171] env[61906]: DEBUG nova.network.neutron [req-8b5d6d30-5a53-4372-b756-8558e6e211ec req-a7ebf362-86a4-4bf5-868c-99dca8df8c4f service nova] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Updated VIF entry in instance network info cache for port b0492e76-fbe6-486e-8fe8-539b1e22ab50. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 919.916343] env[61906]: DEBUG nova.network.neutron [req-8b5d6d30-5a53-4372-b756-8558e6e211ec req-a7ebf362-86a4-4bf5-868c-99dca8df8c4f service nova] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Updating instance_info_cache with network_info: [{"id": "b0492e76-fbe6-486e-8fe8-539b1e22ab50", "address": "fa:16:3e:0f:03:18", "network": {"id": "bc9185e9-f26b-489c-84ee-e6db41ac6283", "bridge": "br-int", "label": "tempest-ServersTestJSON-1167747148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53644a4143a24759a9ff2b5e28b84fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0492e76-fb", "ovs_interfaceid": "b0492e76-fbe6-486e-8fe8-539b1e22ab50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.040541] env[61906]: DEBUG oslo_concurrency.lockutils [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "refresh_cache-ad2db007-e7aa-4c66-9020-ac4acfcff25a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.040789] env[61906]: DEBUG oslo_concurrency.lockutils [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquired lock "refresh_cache-ad2db007-e7aa-4c66-9020-ac4acfcff25a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.040853] env[61906]: DEBUG nova.network.neutron [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 920.048291] env[61906]: DEBUG nova.compute.utils [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 920.050219] env[61906]: DEBUG nova.compute.manager [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 920.050460] env[61906]: DEBUG nova.network.neutron [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 920.092025] env[61906]: DEBUG nova.policy [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '27ba0d3b00ce4513a03259970a600190', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b7d3aa88849f4492aa16ff337b8cfdda', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 920.105342] env[61906]: DEBUG nova.compute.manager [req-babd6766-ee3c-44b1-b1da-a0a01e1a2240 req-ebeae31e-e8a5-43ab-9775-ece885a0fac2 service nova] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Received event network-vif-plugged-d6afef85-5b45-461f-9671-7a187b8022fc {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 920.105807] env[61906]: DEBUG oslo_concurrency.lockutils [req-babd6766-ee3c-44b1-b1da-a0a01e1a2240 req-ebeae31e-e8a5-43ab-9775-ece885a0fac2 service nova] Acquiring lock "ad2db007-e7aa-4c66-9020-ac4acfcff25a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.105807] env[61906]: DEBUG oslo_concurrency.lockutils [req-babd6766-ee3c-44b1-b1da-a0a01e1a2240 req-ebeae31e-e8a5-43ab-9775-ece885a0fac2 service nova] Lock "ad2db007-e7aa-4c66-9020-ac4acfcff25a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.106409] env[61906]: DEBUG oslo_concurrency.lockutils [req-babd6766-ee3c-44b1-b1da-a0a01e1a2240 req-ebeae31e-e8a5-43ab-9775-ece885a0fac2 service nova] Lock "ad2db007-e7aa-4c66-9020-ac4acfcff25a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.106409] env[61906]: DEBUG nova.compute.manager [req-babd6766-ee3c-44b1-b1da-a0a01e1a2240 req-ebeae31e-e8a5-43ab-9775-ece885a0fac2 service nova] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] No waiting events found dispatching network-vif-plugged-d6afef85-5b45-461f-9671-7a187b8022fc {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 920.106409] env[61906]: WARNING nova.compute.manager [req-babd6766-ee3c-44b1-b1da-a0a01e1a2240 req-ebeae31e-e8a5-43ab-9775-ece885a0fac2 service nova] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Received unexpected event network-vif-plugged-d6afef85-5b45-461f-9671-7a187b8022fc for instance with vm_state building and task_state spawning. [ 920.106570] env[61906]: DEBUG nova.compute.manager [req-babd6766-ee3c-44b1-b1da-a0a01e1a2240 req-ebeae31e-e8a5-43ab-9775-ece885a0fac2 service nova] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Received event network-changed-d6afef85-5b45-461f-9671-7a187b8022fc {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 920.106602] env[61906]: DEBUG nova.compute.manager [req-babd6766-ee3c-44b1-b1da-a0a01e1a2240 req-ebeae31e-e8a5-43ab-9775-ece885a0fac2 service nova] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Refreshing instance network info cache due to event network-changed-d6afef85-5b45-461f-9671-7a187b8022fc. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 920.106765] env[61906]: DEBUG oslo_concurrency.lockutils [req-babd6766-ee3c-44b1-b1da-a0a01e1a2240 req-ebeae31e-e8a5-43ab-9775-ece885a0fac2 service nova] Acquiring lock "refresh_cache-ad2db007-e7aa-4c66-9020-ac4acfcff25a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.225596] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5268acfc-6be7-2234-c2b0-5aa0ac99f8f1, 'name': SearchDatastore_Task, 'duration_secs': 0.009339} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.228092] env[61906]: DEBUG oslo_concurrency.lockutils [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.230019] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 920.230019] env[61906]: DEBUG oslo_concurrency.lockutils [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.230019] env[61906]: DEBUG oslo_concurrency.lockutils [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.230019] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 920.230019] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0b4aa3ab-746d-46a9-86a1-0cc10e412ddc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.238341] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 920.238520] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 920.241348] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79805318-c6d8-468e-bec5-f61492d144a2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.247486] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 920.247486] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5278dc29-cafe-2741-c368-a3870c0ecfb0" [ 920.247486] env[61906]: _type = "Task" [ 920.247486] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.255428] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5278dc29-cafe-2741-c368-a3870c0ecfb0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.268179] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0da8c379-15b3-4087-a93d-5ed2cae307c4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.275418] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cccafba-cfd9-4506-91a0-3a6111bcca70 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.304574] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4428774d-60f1-410f-b7a3-9f7bd9d5a7a1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.311831] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-063d4b1a-fd55-4be7-a1b3-2ef5a52a864e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.328261] env[61906]: DEBUG nova.compute.provider_tree [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.340900] env[61906]: DEBUG nova.network.neutron [-] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.363265] env[61906]: DEBUG nova.network.neutron [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Successfully created port: a497b774-60ea-485d-a564-ba7b978b560e {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 920.418655] env[61906]: DEBUG oslo_concurrency.lockutils [req-8b5d6d30-5a53-4372-b756-8558e6e211ec req-a7ebf362-86a4-4bf5-868c-99dca8df8c4f service nova] Releasing lock "refresh_cache-0488c6eb-b510-4d33-ab9f-ebfdab63fb3a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.553523] env[61906]: DEBUG nova.compute.manager [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 920.571106] env[61906]: DEBUG nova.network.neutron [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 920.710901] env[61906]: DEBUG nova.network.neutron [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Updating instance_info_cache with network_info: [{"id": "d6afef85-5b45-461f-9671-7a187b8022fc", "address": "fa:16:3e:73:80:cb", "network": {"id": "d123fdf4-6bb4-400b-a0bb-9d477672af2c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2099881656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d2198383d74dac90aa9727a62176df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6afef85-5b", "ovs_interfaceid": "d6afef85-5b45-461f-9671-7a187b8022fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.757798] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5278dc29-cafe-2741-c368-a3870c0ecfb0, 'name': SearchDatastore_Task, 'duration_secs': 0.008606} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.759986] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b55d13a-eac5-4a1b-b2f3-84cbc9fea9f2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.765263] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 920.765263] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5235e6ba-05d6-3a83-1b5b-2ab84a9edd23" [ 920.765263] env[61906]: _type = "Task" [ 920.765263] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.773491] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5235e6ba-05d6-3a83-1b5b-2ab84a9edd23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.833848] env[61906]: DEBUG nova.scheduler.client.report [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 920.843052] env[61906]: INFO nova.compute.manager [-] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Took 1.27 seconds to deallocate network for instance. [ 921.213449] env[61906]: DEBUG oslo_concurrency.lockutils [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Releasing lock "refresh_cache-ad2db007-e7aa-4c66-9020-ac4acfcff25a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.213809] env[61906]: DEBUG nova.compute.manager [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Instance network_info: |[{"id": "d6afef85-5b45-461f-9671-7a187b8022fc", "address": "fa:16:3e:73:80:cb", "network": {"id": "d123fdf4-6bb4-400b-a0bb-9d477672af2c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2099881656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d2198383d74dac90aa9727a62176df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6afef85-5b", "ovs_interfaceid": "d6afef85-5b45-461f-9671-7a187b8022fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 921.214154] env[61906]: DEBUG oslo_concurrency.lockutils [req-babd6766-ee3c-44b1-b1da-a0a01e1a2240 req-ebeae31e-e8a5-43ab-9775-ece885a0fac2 service nova] Acquired lock "refresh_cache-ad2db007-e7aa-4c66-9020-ac4acfcff25a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.214348] env[61906]: DEBUG nova.network.neutron [req-babd6766-ee3c-44b1-b1da-a0a01e1a2240 req-ebeae31e-e8a5-43ab-9775-ece885a0fac2 service nova] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Refreshing network info cache for port d6afef85-5b45-461f-9671-7a187b8022fc {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 921.215532] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:73:80:cb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cba18f15-a919-422e-a423-1e705e233389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd6afef85-5b45-461f-9671-7a187b8022fc', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 921.223270] env[61906]: DEBUG oslo.service.loopingcall [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 921.224244] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 921.224479] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f6546325-54ba-4902-912c-2f798790350a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.244861] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 921.244861] env[61906]: value = "task-1333347" [ 921.244861] env[61906]: _type = "Task" [ 921.244861] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.256645] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333347, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.275015] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5235e6ba-05d6-3a83-1b5b-2ab84a9edd23, 'name': SearchDatastore_Task, 'duration_secs': 0.009956} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.275314] env[61906]: DEBUG oslo_concurrency.lockutils [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.275568] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a/0488c6eb-b510-4d33-ab9f-ebfdab63fb3a.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 921.275834] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-72c95aa8-1522-4597-ba00-6e27dd09d1ae {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.282430] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 921.282430] env[61906]: value = "task-1333348" [ 921.282430] env[61906]: _type = "Task" [ 921.282430] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.289701] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333348, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.340383] env[61906]: DEBUG oslo_concurrency.lockutils [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.798s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.343022] env[61906]: DEBUG oslo_concurrency.lockutils [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.330s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.344728] env[61906]: INFO nova.compute.claims [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 921.349535] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.362185] env[61906]: INFO nova.scheduler.client.report [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Deleted allocations for instance cc8870d0-eafb-4f73-aa79-b98f51370237 [ 921.385049] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 921.385904] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 921.565020] env[61906]: DEBUG nova.compute.manager [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 921.594529] env[61906]: DEBUG nova.virt.hardware [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 921.594958] env[61906]: DEBUG nova.virt.hardware [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 921.595243] env[61906]: DEBUG nova.virt.hardware [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 921.595608] env[61906]: DEBUG nova.virt.hardware [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 921.595858] env[61906]: DEBUG nova.virt.hardware [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 921.596134] env[61906]: DEBUG nova.virt.hardware [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 921.596540] env[61906]: DEBUG nova.virt.hardware [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 921.596729] env[61906]: DEBUG nova.virt.hardware [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 921.596834] env[61906]: DEBUG nova.virt.hardware [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 921.597040] env[61906]: DEBUG nova.virt.hardware [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 921.597229] env[61906]: DEBUG nova.virt.hardware [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 921.598176] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3373c518-d947-4dd8-b781-ce89cbd80c92 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.608368] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19eab3d3-fc12-4af6-a0bc-43ba18363014 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.754900] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333347, 'name': CreateVM_Task, 'duration_secs': 0.420242} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.755096] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 921.755772] env[61906]: DEBUG oslo_concurrency.lockutils [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.755942] env[61906]: DEBUG oslo_concurrency.lockutils [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.756280] env[61906]: DEBUG oslo_concurrency.lockutils [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 921.756539] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6bf624c0-47c1-45b6-80c3-39f6ed19fa87 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.761593] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 921.761593] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52cab022-149c-634f-8656-a36023f72686" [ 921.761593] env[61906]: _type = "Task" [ 921.761593] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.771657] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52cab022-149c-634f-8656-a36023f72686, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.791886] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333348, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.869618] env[61906]: DEBUG oslo_concurrency.lockutils [None req-067f2563-0675-4958-9247-a9421258e21f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "cc8870d0-eafb-4f73-aa79-b98f51370237" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.732s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.929890] env[61906]: DEBUG nova.compute.manager [req-c8392888-4673-44d0-af76-f3a2e9fc6b11 req-87b2666f-1d5a-47a3-bde5-32935796b78b service nova] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Received event network-vif-plugged-a497b774-60ea-485d-a564-ba7b978b560e {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.931673] env[61906]: DEBUG oslo_concurrency.lockutils [req-c8392888-4673-44d0-af76-f3a2e9fc6b11 req-87b2666f-1d5a-47a3-bde5-32935796b78b service nova] Acquiring lock "1a335019-915b-4065-b22f-d2f0af728d6f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.932441] env[61906]: DEBUG oslo_concurrency.lockutils [req-c8392888-4673-44d0-af76-f3a2e9fc6b11 req-87b2666f-1d5a-47a3-bde5-32935796b78b service nova] Lock "1a335019-915b-4065-b22f-d2f0af728d6f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.932441] env[61906]: DEBUG oslo_concurrency.lockutils [req-c8392888-4673-44d0-af76-f3a2e9fc6b11 req-87b2666f-1d5a-47a3-bde5-32935796b78b service nova] Lock "1a335019-915b-4065-b22f-d2f0af728d6f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.932441] env[61906]: DEBUG nova.compute.manager [req-c8392888-4673-44d0-af76-f3a2e9fc6b11 req-87b2666f-1d5a-47a3-bde5-32935796b78b service nova] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] No waiting events found dispatching network-vif-plugged-a497b774-60ea-485d-a564-ba7b978b560e {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 921.932751] env[61906]: WARNING nova.compute.manager [req-c8392888-4673-44d0-af76-f3a2e9fc6b11 req-87b2666f-1d5a-47a3-bde5-32935796b78b service nova] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Received unexpected event network-vif-plugged-a497b774-60ea-485d-a564-ba7b978b560e for instance with vm_state building and task_state spawning. [ 921.969465] env[61906]: DEBUG nova.network.neutron [req-babd6766-ee3c-44b1-b1da-a0a01e1a2240 req-ebeae31e-e8a5-43ab-9775-ece885a0fac2 service nova] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Updated VIF entry in instance network info cache for port d6afef85-5b45-461f-9671-7a187b8022fc. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 921.969832] env[61906]: DEBUG nova.network.neutron [req-babd6766-ee3c-44b1-b1da-a0a01e1a2240 req-ebeae31e-e8a5-43ab-9775-ece885a0fac2 service nova] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Updating instance_info_cache with network_info: [{"id": "d6afef85-5b45-461f-9671-7a187b8022fc", "address": "fa:16:3e:73:80:cb", "network": {"id": "d123fdf4-6bb4-400b-a0bb-9d477672af2c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2099881656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d2198383d74dac90aa9727a62176df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6afef85-5b", "ovs_interfaceid": "d6afef85-5b45-461f-9671-7a187b8022fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.017104] env[61906]: DEBUG nova.network.neutron [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Successfully updated port: a497b774-60ea-485d-a564-ba7b978b560e {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 922.273336] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52cab022-149c-634f-8656-a36023f72686, 'name': SearchDatastore_Task, 'duration_secs': 0.008798} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.273691] env[61906]: DEBUG oslo_concurrency.lockutils [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.273917] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 922.274176] env[61906]: DEBUG oslo_concurrency.lockutils [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.274329] env[61906]: DEBUG oslo_concurrency.lockutils [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.274518] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 922.274788] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9fec6fe3-597f-4c96-adb6-a4c527ea5851 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.290672] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 922.290672] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 922.293720] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f3d8b33-ee68-42e2-8bed-534f4681c44d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.295821] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333348, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.299148] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 922.299148] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]521957f4-0897-26c3-8e4c-72146119a265" [ 922.299148] env[61906]: _type = "Task" [ 922.299148] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.306683] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]521957f4-0897-26c3-8e4c-72146119a265, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.471979] env[61906]: DEBUG oslo_concurrency.lockutils [req-babd6766-ee3c-44b1-b1da-a0a01e1a2240 req-ebeae31e-e8a5-43ab-9775-ece885a0fac2 service nova] Releasing lock "refresh_cache-ad2db007-e7aa-4c66-9020-ac4acfcff25a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.472351] env[61906]: DEBUG nova.compute.manager [req-babd6766-ee3c-44b1-b1da-a0a01e1a2240 req-ebeae31e-e8a5-43ab-9775-ece885a0fac2 service nova] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Received event network-vif-deleted-531cc05e-2f49-4b47-84da-bef5affc1bf2 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 922.472586] env[61906]: INFO nova.compute.manager [req-babd6766-ee3c-44b1-b1da-a0a01e1a2240 req-ebeae31e-e8a5-43ab-9775-ece885a0fac2 service nova] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Neutron deleted interface 531cc05e-2f49-4b47-84da-bef5affc1bf2; detaching it from the instance and deleting it from the info cache [ 922.472802] env[61906]: DEBUG nova.network.neutron [req-babd6766-ee3c-44b1-b1da-a0a01e1a2240 req-ebeae31e-e8a5-43ab-9775-ece885a0fac2 service nova] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.519203] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.519391] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.519568] env[61906]: DEBUG nova.network.neutron [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 922.526286] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b71c539-b8d7-495e-9baf-2ef9cef591cf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.537716] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a0869a-5722-48e3-978b-c1fa297836b9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.570882] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c9f6b58-9597-4691-a61d-2c1113f70582 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.578587] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de55c3a0-4a47-4f13-bd2d-792b5a948730 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.592993] env[61906]: DEBUG nova.compute.provider_tree [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.794298] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333348, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.475616} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.794610] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a/0488c6eb-b510-4d33-ab9f-ebfdab63fb3a.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 922.794869] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 922.795154] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f21d6e01-46f1-431b-a2e2-a1c5f8afbe3e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.804397] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 922.804397] env[61906]: value = "task-1333349" [ 922.804397] env[61906]: _type = "Task" [ 922.804397] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.810537] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]521957f4-0897-26c3-8e4c-72146119a265, 'name': SearchDatastore_Task, 'duration_secs': 0.068996} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.811600] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d867fa0-5c81-4abd-9ba5-c5dc79963fa0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.816526] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333349, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.819887] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 922.819887] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52a1608d-a74f-4892-44a9-61ddc799f03d" [ 922.819887] env[61906]: _type = "Task" [ 922.819887] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.827818] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52a1608d-a74f-4892-44a9-61ddc799f03d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.975511] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-67b97123-6fbb-4e74-ad38-e433606d9dc5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.989680] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c04c6319-7d5f-4b0a-a4ae-cf6c166191a6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.021261] env[61906]: DEBUG nova.compute.manager [req-babd6766-ee3c-44b1-b1da-a0a01e1a2240 req-ebeae31e-e8a5-43ab-9775-ece885a0fac2 service nova] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Detach interface failed, port_id=531cc05e-2f49-4b47-84da-bef5affc1bf2, reason: Instance d60575a6-af13-4d81-95d1-081d0f3e2ce6 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 923.050228] env[61906]: DEBUG nova.network.neutron [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 923.096247] env[61906]: DEBUG nova.scheduler.client.report [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 923.116306] env[61906]: DEBUG oslo_concurrency.lockutils [None req-93e3c3cd-93f8-4bab-bcb7-a66ce20a5655 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "interface-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.116603] env[61906]: DEBUG oslo_concurrency.lockutils [None req-93e3c3cd-93f8-4bab-bcb7-a66ce20a5655 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "interface-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.116928] env[61906]: DEBUG nova.objects.instance [None req-93e3c3cd-93f8-4bab-bcb7-a66ce20a5655 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lazy-loading 'flavor' on Instance uuid a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.252218] env[61906]: DEBUG nova.network.neutron [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Updating instance_info_cache with network_info: [{"id": "a497b774-60ea-485d-a564-ba7b978b560e", "address": "fa:16:3e:0d:cf:6e", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa497b774-60", "ovs_interfaceid": "a497b774-60ea-485d-a564-ba7b978b560e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.315076] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333349, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.111382} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.315430] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 923.316208] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f3b1a39-3d8d-4214-b142-e6366085ad6f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.339579] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a/0488c6eb-b510-4d33-ab9f-ebfdab63fb3a.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 923.342696] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-85f43811-f2c1-4fce-9436-9d0e68dd57e6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.364308] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52a1608d-a74f-4892-44a9-61ddc799f03d, 'name': SearchDatastore_Task, 'duration_secs': 0.0098} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.365564] env[61906]: DEBUG oslo_concurrency.lockutils [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.365822] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] ad2db007-e7aa-4c66-9020-ac4acfcff25a/ad2db007-e7aa-4c66-9020-ac4acfcff25a.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 923.366164] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 923.366164] env[61906]: value = "task-1333350" [ 923.366164] env[61906]: _type = "Task" [ 923.366164] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.366382] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a0d0fb9c-d097-4106-8101-7cff789c984b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.377926] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333350, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.379643] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 923.379643] env[61906]: value = "task-1333351" [ 923.379643] env[61906]: _type = "Task" [ 923.379643] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.389244] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333351, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.601955] env[61906]: DEBUG oslo_concurrency.lockutils [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.259s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.602693] env[61906]: DEBUG nova.compute.manager [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 923.606092] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.330s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.607687] env[61906]: INFO nova.compute.claims [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 923.621345] env[61906]: DEBUG nova.objects.instance [None req-93e3c3cd-93f8-4bab-bcb7-a66ce20a5655 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lazy-loading 'pci_requests' on Instance uuid a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.758906] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Releasing lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.759325] env[61906]: DEBUG nova.compute.manager [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Instance network_info: |[{"id": "a497b774-60ea-485d-a564-ba7b978b560e", "address": "fa:16:3e:0d:cf:6e", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa497b774-60", "ovs_interfaceid": "a497b774-60ea-485d-a564-ba7b978b560e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 923.759782] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0d:cf:6e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a497b774-60ea-485d-a564-ba7b978b560e', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 923.768159] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Creating folder: Project (b7d3aa88849f4492aa16ff337b8cfdda). Parent ref: group-v284713. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 923.768540] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bed5db5d-df21-4c92-aa20-3e90dc08ed39 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.781909] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Created folder: Project (b7d3aa88849f4492aa16ff337b8cfdda) in parent group-v284713. [ 923.782139] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Creating folder: Instances. Parent ref: group-v284789. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 923.782542] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-431bcd76-7f70-4561-a2cb-2c546fa28405 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.794037] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Created folder: Instances in parent group-v284789. [ 923.794339] env[61906]: DEBUG oslo.service.loopingcall [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 923.794658] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 923.794896] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3d7db870-4176-493c-abc9-5c6c1ee7305e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.816763] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 923.816763] env[61906]: value = "task-1333354" [ 923.816763] env[61906]: _type = "Task" [ 923.816763] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.827781] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333354, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.879273] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333350, 'name': ReconfigVM_Task, 'duration_secs': 0.448127} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.879746] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Reconfigured VM instance instance-0000004c to attach disk [datastore1] 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a/0488c6eb-b510-4d33-ab9f-ebfdab63fb3a.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 923.880727] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b7fcedb4-ae34-48d6-ab2f-259e261ace6f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.895676] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333351, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.469776} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.897147] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] ad2db007-e7aa-4c66-9020-ac4acfcff25a/ad2db007-e7aa-4c66-9020-ac4acfcff25a.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 923.897412] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 923.899089] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 923.899089] env[61906]: value = "task-1333355" [ 923.899089] env[61906]: _type = "Task" [ 923.899089] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.900051] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-44755f87-4dff-4a9d-9b59-be027c65751e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.910958] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333355, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.912141] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 923.912141] env[61906]: value = "task-1333356" [ 923.912141] env[61906]: _type = "Task" [ 923.912141] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.921273] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333356, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.040266] env[61906]: DEBUG nova.compute.manager [req-848d2b02-510e-42f5-94e4-0965c3066ac0 req-68aafef2-5a2e-4f82-86d1-4eb30b08b64c service nova] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Received event network-changed-a497b774-60ea-485d-a564-ba7b978b560e {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 924.040572] env[61906]: DEBUG nova.compute.manager [req-848d2b02-510e-42f5-94e4-0965c3066ac0 req-68aafef2-5a2e-4f82-86d1-4eb30b08b64c service nova] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Refreshing instance network info cache due to event network-changed-a497b774-60ea-485d-a564-ba7b978b560e. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 924.040893] env[61906]: DEBUG oslo_concurrency.lockutils [req-848d2b02-510e-42f5-94e4-0965c3066ac0 req-68aafef2-5a2e-4f82-86d1-4eb30b08b64c service nova] Acquiring lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.041174] env[61906]: DEBUG oslo_concurrency.lockutils [req-848d2b02-510e-42f5-94e4-0965c3066ac0 req-68aafef2-5a2e-4f82-86d1-4eb30b08b64c service nova] Acquired lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.041455] env[61906]: DEBUG nova.network.neutron [req-848d2b02-510e-42f5-94e4-0965c3066ac0 req-68aafef2-5a2e-4f82-86d1-4eb30b08b64c service nova] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Refreshing network info cache for port a497b774-60ea-485d-a564-ba7b978b560e {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 924.112798] env[61906]: DEBUG nova.compute.utils [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 924.116824] env[61906]: DEBUG nova.compute.manager [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Not allocating networking since 'none' was specified. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 924.124765] env[61906]: DEBUG nova.objects.base [None req-93e3c3cd-93f8-4bab-bcb7-a66ce20a5655 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 924.124765] env[61906]: DEBUG nova.network.neutron [None req-93e3c3cd-93f8-4bab-bcb7-a66ce20a5655 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 924.217991] env[61906]: DEBUG oslo_concurrency.lockutils [None req-93e3c3cd-93f8-4bab-bcb7-a66ce20a5655 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "interface-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.101s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.327620] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333354, 'name': CreateVM_Task, 'duration_secs': 0.399502} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.327959] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 924.328476] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.328648] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.328972] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 924.329249] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e0623d4-cf2a-4e0d-959a-6685459a4de2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.334737] env[61906]: DEBUG oslo_vmware.api [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 924.334737] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52c7a355-585f-90ec-560a-1f9d583adc91" [ 924.334737] env[61906]: _type = "Task" [ 924.334737] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.343246] env[61906]: DEBUG oslo_vmware.api [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52c7a355-585f-90ec-560a-1f9d583adc91, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.385534] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 924.385928] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Starting heal instance info cache {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 924.411962] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333355, 'name': Rename_Task, 'duration_secs': 0.175478} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.412296] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 924.412622] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-15aab576-d5a1-42c0-8ecf-263d73564a90 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.423430] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333356, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078134} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.424704] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 924.425076] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 924.425076] env[61906]: value = "task-1333357" [ 924.425076] env[61906]: _type = "Task" [ 924.425076] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.425777] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b91caf0-214f-4322-b494-46916001dbd6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.452055] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] ad2db007-e7aa-4c66-9020-ac4acfcff25a/ad2db007-e7aa-4c66-9020-ac4acfcff25a.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 924.454956] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9e071b3-c6a8-4b8c-b78a-e592040501c9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.469755] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333357, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.476902] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 924.476902] env[61906]: value = "task-1333358" [ 924.476902] env[61906]: _type = "Task" [ 924.476902] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.490577] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333358, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.617533] env[61906]: DEBUG nova.compute.manager [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 924.804068] env[61906]: DEBUG nova.network.neutron [req-848d2b02-510e-42f5-94e4-0965c3066ac0 req-68aafef2-5a2e-4f82-86d1-4eb30b08b64c service nova] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Updated VIF entry in instance network info cache for port a497b774-60ea-485d-a564-ba7b978b560e. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 924.804609] env[61906]: DEBUG nova.network.neutron [req-848d2b02-510e-42f5-94e4-0965c3066ac0 req-68aafef2-5a2e-4f82-86d1-4eb30b08b64c service nova] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Updating instance_info_cache with network_info: [{"id": "a497b774-60ea-485d-a564-ba7b978b560e", "address": "fa:16:3e:0d:cf:6e", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa497b774-60", "ovs_interfaceid": "a497b774-60ea-485d-a564-ba7b978b560e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.844932] env[61906]: DEBUG oslo_vmware.api [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52c7a355-585f-90ec-560a-1f9d583adc91, 'name': SearchDatastore_Task, 'duration_secs': 0.015443} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.845262] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.845510] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 924.845746] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.845896] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.846095] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 924.847143] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a95b17cd-ed10-41eb-808b-7b6a0dd961a5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.849377] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c14ac44-5d42-465c-a151-e2489428640d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.857654] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-931e4d3d-13b8-43f7-a086-9a74a97b84bd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.862108] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 924.862315] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 924.863545] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e5280f3-26d2-42f1-9e89-f4ee6021ecba {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.890076] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f310ffdc-398a-4684-aab0-e617048f648c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.895378] env[61906]: DEBUG oslo_vmware.api [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 924.895378] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52e3d988-e952-27b5-7a6e-41917da30304" [ 924.895378] env[61906]: _type = "Task" [ 924.895378] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.902019] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a995e498-9b9e-4b89-be99-6c456480463e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.908970] env[61906]: DEBUG oslo_vmware.api [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52e3d988-e952-27b5-7a6e-41917da30304, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.921572] env[61906]: DEBUG nova.compute.provider_tree [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 924.938698] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333357, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.988788] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333358, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.309707] env[61906]: DEBUG oslo_concurrency.lockutils [req-848d2b02-510e-42f5-94e4-0965c3066ac0 req-68aafef2-5a2e-4f82-86d1-4eb30b08b64c service nova] Releasing lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.409571] env[61906]: DEBUG oslo_vmware.api [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52e3d988-e952-27b5-7a6e-41917da30304, 'name': SearchDatastore_Task, 'duration_secs': 0.018006} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.410856] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32b8c896-9efc-4da0-9063-091fb9a7ea24 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.416703] env[61906]: DEBUG oslo_vmware.api [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 925.416703] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]528f9907-8a4d-0848-0fb3-c83167a54796" [ 925.416703] env[61906]: _type = "Task" [ 925.416703] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.420162] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "refresh_cache-d60575a6-af13-4d81-95d1-081d0f3e2ce6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.420303] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquired lock "refresh_cache-d60575a6-af13-4d81-95d1-081d0f3e2ce6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.420446] env[61906]: DEBUG nova.network.neutron [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Forcefully refreshing network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 925.425030] env[61906]: DEBUG nova.scheduler.client.report [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 925.431992] env[61906]: DEBUG oslo_vmware.api [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]528f9907-8a4d-0848-0fb3-c83167a54796, 'name': SearchDatastore_Task, 'duration_secs': 0.009187} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.435284] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.435546] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 1a335019-915b-4065-b22f-d2f0af728d6f/1a335019-915b-4065-b22f-d2f0af728d6f.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 925.436056] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-51865c9f-d82a-4f93-86e4-2f687912a7d4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.443849] env[61906]: DEBUG oslo_vmware.api [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333357, 'name': PowerOnVM_Task, 'duration_secs': 0.698439} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.445013] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 925.445234] env[61906]: INFO nova.compute.manager [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Took 8.49 seconds to spawn the instance on the hypervisor. [ 925.445417] env[61906]: DEBUG nova.compute.manager [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 925.445738] env[61906]: DEBUG oslo_vmware.api [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 925.445738] env[61906]: value = "task-1333359" [ 925.445738] env[61906]: _type = "Task" [ 925.445738] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.446428] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9924912e-9bb1-44be-a481-96b227a2048c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.488899] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333358, 'name': ReconfigVM_Task, 'duration_secs': 0.595742} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.489205] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Reconfigured VM instance instance-0000004d to attach disk [datastore1] ad2db007-e7aa-4c66-9020-ac4acfcff25a/ad2db007-e7aa-4c66-9020-ac4acfcff25a.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 925.490575] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-83afcaf7-7281-4aed-9aa9-0ab5ed3b7e3b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.497308] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 925.497308] env[61906]: value = "task-1333360" [ 925.497308] env[61906]: _type = "Task" [ 925.497308] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.506635] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333360, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.631496] env[61906]: DEBUG nova.compute.manager [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 925.657914] env[61906]: DEBUG nova.virt.hardware [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 925.658239] env[61906]: DEBUG nova.virt.hardware [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 925.658438] env[61906]: DEBUG nova.virt.hardware [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 925.658653] env[61906]: DEBUG nova.virt.hardware [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 925.658845] env[61906]: DEBUG nova.virt.hardware [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 925.659049] env[61906]: DEBUG nova.virt.hardware [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 925.659348] env[61906]: DEBUG nova.virt.hardware [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 925.659570] env[61906]: DEBUG nova.virt.hardware [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 925.659830] env[61906]: DEBUG nova.virt.hardware [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 925.660082] env[61906]: DEBUG nova.virt.hardware [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 925.660655] env[61906]: DEBUG nova.virt.hardware [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 925.661378] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4588ec96-a2a0-49b0-8ca6-c463c1a51db6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.670528] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3296fa3-fce1-46c4-979b-2669356a0780 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.687628] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Instance VIF info [] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 925.693497] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Creating folder: Project (bfdd608c6bf944299e64e30f2c85f210). Parent ref: group-v284713. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 925.694977] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1f0c59f6-0d23-48f0-bf55-bb3fb14ac18e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.710729] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Created folder: Project (bfdd608c6bf944299e64e30f2c85f210) in parent group-v284713. [ 925.711013] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Creating folder: Instances. Parent ref: group-v284792. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 925.712031] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8e2150f9-4d6b-446e-9b25-0f29c404c907 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.725286] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Created folder: Instances in parent group-v284792. [ 925.725643] env[61906]: DEBUG oslo.service.loopingcall [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 925.725868] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 925.726128] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dde63b9a-f165-427c-ab9c-5bda577223d1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.745043] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 925.745043] env[61906]: value = "task-1333363" [ 925.745043] env[61906]: _type = "Task" [ 925.745043] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.754498] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333363, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.934102] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.328s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.934686] env[61906]: DEBUG nova.compute.manager [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 925.937324] env[61906]: DEBUG oslo_concurrency.lockutils [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.286s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.937557] env[61906]: DEBUG nova.objects.instance [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Lazy-loading 'resources' on Instance uuid 0e86bfdf-9002-4a2d-a2ac-af35921b020e {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.940506] env[61906]: DEBUG nova.network.neutron [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 925.965998] env[61906]: DEBUG oslo_vmware.api [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333359, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.968399] env[61906]: INFO nova.compute.manager [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Took 13.34 seconds to build instance. [ 926.007295] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333360, 'name': Rename_Task, 'duration_secs': 0.165089} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.007584] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 926.007878] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4b3d97db-d33a-480c-a3a0-f6add501ab11 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.015952] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 926.015952] env[61906]: value = "task-1333364" [ 926.015952] env[61906]: _type = "Task" [ 926.015952] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.038034] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333364, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.200744] env[61906]: DEBUG oslo_concurrency.lockutils [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "interface-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.200872] env[61906]: DEBUG oslo_concurrency.lockutils [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "interface-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.201959] env[61906]: DEBUG nova.objects.instance [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lazy-loading 'flavor' on Instance uuid a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.255522] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333363, 'name': CreateVM_Task, 'duration_secs': 0.321058} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.255703] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 926.256152] env[61906]: DEBUG oslo_concurrency.lockutils [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.256323] env[61906]: DEBUG oslo_concurrency.lockutils [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.256664] env[61906]: DEBUG oslo_concurrency.lockutils [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 926.256926] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfb7312a-70f7-4983-b7be-c48a054d4e40 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.262241] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 926.262241] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52f0faea-dfa2-af20-2d4c-359d08e72d5a" [ 926.262241] env[61906]: _type = "Task" [ 926.262241] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.270241] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52f0faea-dfa2-af20-2d4c-359d08e72d5a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.441278] env[61906]: DEBUG nova.compute.utils [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 926.442815] env[61906]: DEBUG nova.compute.manager [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Not allocating networking since 'none' was specified. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 926.461980] env[61906]: DEBUG oslo_vmware.api [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333359, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51964} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.464580] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 1a335019-915b-4065-b22f-d2f0af728d6f/1a335019-915b-4065-b22f-d2f0af728d6f.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 926.464817] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 926.465873] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4bf888b8-3361-4ea3-a2ab-6d15a754ec87 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.470402] env[61906]: DEBUG oslo_concurrency.lockutils [None req-314efc45-ea8a-461e-9492-0b7b2298a46e tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "0488c6eb-b510-4d33-ab9f-ebfdab63fb3a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.851s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.473579] env[61906]: DEBUG oslo_vmware.api [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 926.473579] env[61906]: value = "task-1333365" [ 926.473579] env[61906]: _type = "Task" [ 926.473579] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.485696] env[61906]: DEBUG oslo_vmware.api [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333365, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.527213] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333364, 'name': PowerOnVM_Task} progress is 19%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.530590] env[61906]: DEBUG nova.network.neutron [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.636115] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-046d76e6-adc0-4a00-82de-be8b44380941 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.644637] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b767975-3380-4a4f-97fe-c2b2a2b48641 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.674258] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abebbacb-9f29-4f65-bb1c-7859d6dc7c9f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.682166] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10fa5f33-9042-4054-8640-75d2cdebd3ea {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.695784] env[61906]: DEBUG nova.compute.provider_tree [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.773397] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52f0faea-dfa2-af20-2d4c-359d08e72d5a, 'name': SearchDatastore_Task, 'duration_secs': 0.009049} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.773734] env[61906]: DEBUG oslo_concurrency.lockutils [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.773974] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 926.774227] env[61906]: DEBUG oslo_concurrency.lockutils [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.774380] env[61906]: DEBUG oslo_concurrency.lockutils [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.774565] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 926.774833] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0f140fdd-272f-4041-b101-3ef4a921db22 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.777929] env[61906]: DEBUG nova.objects.instance [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lazy-loading 'pci_requests' on Instance uuid a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.795345] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 926.795542] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 926.796314] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2995a76b-4a3d-4846-85d0-cd1710680340 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.802070] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 926.802070] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5281f859-275f-6125-1b84-03cd584efae2" [ 926.802070] env[61906]: _type = "Task" [ 926.802070] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.810553] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5281f859-275f-6125-1b84-03cd584efae2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.948619] env[61906]: DEBUG nova.compute.manager [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 926.983584] env[61906]: DEBUG oslo_vmware.api [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333365, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.249734} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.983869] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 926.984651] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a13f4c-e7c0-48cc-83d9-2d15a4a45fa4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.007909] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 1a335019-915b-4065-b22f-d2f0af728d6f/1a335019-915b-4065-b22f-d2f0af728d6f.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 927.008575] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d61dc5bb-fcad-4537-a4b2-94235a7c832d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.032784] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Releasing lock "refresh_cache-d60575a6-af13-4d81-95d1-081d0f3e2ce6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.032939] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Updated the network info_cache for instance {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 927.033101] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 927.033261] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 927.033420] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 927.033559] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 927.033704] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 927.033833] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61906) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 927.033972] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 927.038912] env[61906]: DEBUG oslo_vmware.api [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 927.038912] env[61906]: value = "task-1333366" [ 927.038912] env[61906]: _type = "Task" [ 927.038912] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.041847] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333364, 'name': PowerOnVM_Task} progress is 91%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.051242] env[61906]: DEBUG oslo_vmware.api [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333366, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.199338] env[61906]: DEBUG nova.scheduler.client.report [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 927.280797] env[61906]: DEBUG nova.objects.base [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 927.281458] env[61906]: DEBUG nova.network.neutron [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 927.313056] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5281f859-275f-6125-1b84-03cd584efae2, 'name': SearchDatastore_Task, 'duration_secs': 0.009763} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.313895] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4bc2fc1f-b381-4274-8dcc-b285d0a5bb2a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.319148] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 927.319148] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]529b4ebd-0c6f-2d9b-f637-ba28498a3f3c" [ 927.319148] env[61906]: _type = "Task" [ 927.319148] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.327466] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]529b4ebd-0c6f-2d9b-f637-ba28498a3f3c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.337693] env[61906]: DEBUG nova.policy [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0de9abfd085f4a43a878da8448615bc4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e36ff7a500ef444cbf5a168c5a48208b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 927.401936] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "0488c6eb-b510-4d33-ab9f-ebfdab63fb3a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.402231] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "0488c6eb-b510-4d33-ab9f-ebfdab63fb3a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.402515] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "0488c6eb-b510-4d33-ab9f-ebfdab63fb3a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.402710] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "0488c6eb-b510-4d33-ab9f-ebfdab63fb3a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.402884] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "0488c6eb-b510-4d33-ab9f-ebfdab63fb3a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.405057] env[61906]: INFO nova.compute.manager [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Terminating instance [ 927.406832] env[61906]: DEBUG nova.compute.manager [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 927.406987] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 927.407809] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0836dca3-8d66-4977-be18-a36f0dc08e57 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.416485] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 927.416716] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5c77c151-16c5-4963-8eb5-928192a32a66 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.423988] env[61906]: DEBUG oslo_vmware.api [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 927.423988] env[61906]: value = "task-1333367" [ 927.423988] env[61906]: _type = "Task" [ 927.423988] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.433153] env[61906]: DEBUG oslo_vmware.api [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333367, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.533623] env[61906]: DEBUG oslo_vmware.api [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333364, 'name': PowerOnVM_Task, 'duration_secs': 1.046438} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.533953] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 927.534564] env[61906]: INFO nova.compute.manager [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Took 8.27 seconds to spawn the instance on the hypervisor. [ 927.534564] env[61906]: DEBUG nova.compute.manager [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 927.535101] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3e91557-ef87-4129-89e5-c88df8168b5d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.538226] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.555216] env[61906]: DEBUG oslo_vmware.api [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333366, 'name': ReconfigVM_Task, 'duration_secs': 0.280661} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.555552] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 1a335019-915b-4065-b22f-d2f0af728d6f/1a335019-915b-4065-b22f-d2f0af728d6f.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 927.556380] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8c211597-8407-427b-b649-d3359b74e036 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.564253] env[61906]: DEBUG oslo_vmware.api [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 927.564253] env[61906]: value = "task-1333368" [ 927.564253] env[61906]: _type = "Task" [ 927.564253] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.574083] env[61906]: DEBUG oslo_vmware.api [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333368, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.614543] env[61906]: DEBUG nova.network.neutron [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Successfully created port: c2232d91-1c2c-4a66-9026-933169da90e1 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 927.704760] env[61906]: DEBUG oslo_concurrency.lockutils [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.767s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.707204] env[61906]: DEBUG oslo_concurrency.lockutils [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.256s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.707422] env[61906]: DEBUG nova.objects.instance [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lazy-loading 'resources' on Instance uuid 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.730405] env[61906]: INFO nova.scheduler.client.report [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Deleted allocations for instance 0e86bfdf-9002-4a2d-a2ac-af35921b020e [ 927.830545] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]529b4ebd-0c6f-2d9b-f637-ba28498a3f3c, 'name': SearchDatastore_Task, 'duration_secs': 0.013429} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.830856] env[61906]: DEBUG oslo_concurrency.lockutils [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.831085] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 11cb9d89-2304-45a6-b4bb-fcf8417a0518/11cb9d89-2304-45a6-b4bb-fcf8417a0518.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 927.831366] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d606bbfd-e39b-42d1-b94f-cb939281c8d0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.840391] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 927.840391] env[61906]: value = "task-1333369" [ 927.840391] env[61906]: _type = "Task" [ 927.840391] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.849013] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333369, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.936824] env[61906]: DEBUG oslo_vmware.api [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333367, 'name': PowerOffVM_Task, 'duration_secs': 0.219477} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.937058] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 927.937273] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 927.937558] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-19f0e4ad-e30d-4f2b-b9a3-e0fc9efc914a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.958075] env[61906]: DEBUG nova.compute.manager [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 927.980739] env[61906]: DEBUG nova.virt.hardware [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 927.980739] env[61906]: DEBUG nova.virt.hardware [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 927.981059] env[61906]: DEBUG nova.virt.hardware [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 927.981348] env[61906]: DEBUG nova.virt.hardware [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 927.981579] env[61906]: DEBUG nova.virt.hardware [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 927.981808] env[61906]: DEBUG nova.virt.hardware [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 927.982096] env[61906]: DEBUG nova.virt.hardware [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 927.982343] env[61906]: DEBUG nova.virt.hardware [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 927.982570] env[61906]: DEBUG nova.virt.hardware [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 927.982775] env[61906]: DEBUG nova.virt.hardware [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 927.982990] env[61906]: DEBUG nova.virt.hardware [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 927.983904] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-982d1b6b-b1e1-49fe-acfd-84b4e6b63272 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.993683] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b7c9f10-cf20-4092-9711-013bcbeca1c9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.999114] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 927.999410] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 927.999646] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Deleting the datastore file [datastore1] 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 927.999964] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fec53f6b-f6f8-434a-95fd-4895419dbeb6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.010037] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Instance VIF info [] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 928.015776] env[61906]: DEBUG oslo.service.loopingcall [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 928.016416] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 928.016679] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-19aa6234-6f36-42a8-9166-0f4f68f86330 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.029187] env[61906]: DEBUG oslo_vmware.api [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 928.029187] env[61906]: value = "task-1333371" [ 928.029187] env[61906]: _type = "Task" [ 928.029187] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.034135] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 928.034135] env[61906]: value = "task-1333372" [ 928.034135] env[61906]: _type = "Task" [ 928.034135] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.039459] env[61906]: DEBUG oslo_vmware.api [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333371, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.044188] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333372, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.056919] env[61906]: INFO nova.compute.manager [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Took 15.11 seconds to build instance. [ 928.082389] env[61906]: DEBUG oslo_vmware.api [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333368, 'name': Rename_Task, 'duration_secs': 0.151286} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.082744] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 928.083034] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b7dc3b36-5482-4969-88b3-8fedef78856f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.090376] env[61906]: DEBUG oslo_vmware.api [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 928.090376] env[61906]: value = "task-1333373" [ 928.090376] env[61906]: _type = "Task" [ 928.090376] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.101433] env[61906]: DEBUG oslo_vmware.api [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333373, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.240072] env[61906]: DEBUG oslo_concurrency.lockutils [None req-326d5287-b26c-48a0-b9e0-c183b3ec0ba2 tempest-ServersNegativeTestMultiTenantJSON-1906172250 tempest-ServersNegativeTestMultiTenantJSON-1906172250-project-member] Lock "0e86bfdf-9002-4a2d-a2ac-af35921b020e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.516s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.349727] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333369, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.427066] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e536a327-ef90-499f-9f4e-3e1d9cfe010d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.435086] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a39d6ff-30be-4dfa-9f48-f06e7d3172fb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.468656] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3ae6c69-8204-41b2-84bf-8e1e254cf634 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.477593] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fd44113-94f3-405a-b133-56e15057771b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.493886] env[61906]: DEBUG nova.compute.provider_tree [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 928.543343] env[61906]: DEBUG oslo_vmware.api [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333371, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.549244] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333372, 'name': CreateVM_Task, 'duration_secs': 0.344534} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.549583] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 928.549880] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.550030] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.550420] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 928.550756] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a040c895-1e3a-4b18-a76f-d7d918fa8494 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.555736] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 928.555736] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]520b4b18-f34e-f8fe-b293-cbe6e52d7542" [ 928.555736] env[61906]: _type = "Task" [ 928.555736] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.559039] env[61906]: DEBUG oslo_concurrency.lockutils [None req-28f73bdd-b4d5-4ce3-a1e6-b79fcd064264 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "ad2db007-e7aa-4c66-9020-ac4acfcff25a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.624s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.564727] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]520b4b18-f34e-f8fe-b293-cbe6e52d7542, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.603588] env[61906]: DEBUG oslo_vmware.api [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333373, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.850657] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333369, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.612328} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.850930] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 11cb9d89-2304-45a6-b4bb-fcf8417a0518/11cb9d89-2304-45a6-b4bb-fcf8417a0518.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 928.851166] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 928.851454] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-03dbe1ef-27e7-4516-899d-b6872355760a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.858676] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 928.858676] env[61906]: value = "task-1333374" [ 928.858676] env[61906]: _type = "Task" [ 928.858676] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.869119] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333374, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.998164] env[61906]: DEBUG nova.scheduler.client.report [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 929.005024] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f53ff00-4dc7-49cf-ab6b-a3b5381124b6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.011240] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-54cca8b8-a2a9-4212-b03b-fd79a7202db3 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Suspending the VM {{(pid=61906) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 929.011240] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-7b3181ee-b8ec-403d-9512-a8c889a08a80 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.020399] env[61906]: DEBUG oslo_vmware.api [None req-54cca8b8-a2a9-4212-b03b-fd79a7202db3 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 929.020399] env[61906]: value = "task-1333375" [ 929.020399] env[61906]: _type = "Task" [ 929.020399] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.030860] env[61906]: DEBUG oslo_vmware.api [None req-54cca8b8-a2a9-4212-b03b-fd79a7202db3 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333375, 'name': SuspendVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.040426] env[61906]: DEBUG oslo_vmware.api [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333371, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.068750] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]520b4b18-f34e-f8fe-b293-cbe6e52d7542, 'name': SearchDatastore_Task, 'duration_secs': 0.013252} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.069413] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.069653] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 929.069949] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.070156] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.070524] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 929.070920] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1ff6d94b-b6b2-4b04-a7ac-68eb3d9ac3ea {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.081398] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 929.081692] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 929.082593] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6e0277a-ac46-48e2-b195-5773740a6db6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.089553] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 929.089553] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]525dbd61-2d2e-98ee-4f38-8c6cdd86ecfb" [ 929.089553] env[61906]: _type = "Task" [ 929.089553] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.102180] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]525dbd61-2d2e-98ee-4f38-8c6cdd86ecfb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.105669] env[61906]: DEBUG oslo_vmware.api [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333373, 'name': PowerOnVM_Task, 'duration_secs': 0.661644} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.106145] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 929.106386] env[61906]: INFO nova.compute.manager [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Took 7.54 seconds to spawn the instance on the hypervisor. [ 929.106613] env[61906]: DEBUG nova.compute.manager [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.107483] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef6925a9-7552-41ea-9c40-dffd1092f4c1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.132217] env[61906]: DEBUG nova.network.neutron [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Successfully updated port: c2232d91-1c2c-4a66-9026-933169da90e1 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 929.143979] env[61906]: DEBUG nova.compute.manager [req-c8394cfe-37f6-4002-a002-5ab639e7318d req-bbc50369-5499-4d5a-bfdd-dad20fe3064d service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Received event network-vif-plugged-c2232d91-1c2c-4a66-9026-933169da90e1 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 929.143979] env[61906]: DEBUG oslo_concurrency.lockutils [req-c8394cfe-37f6-4002-a002-5ab639e7318d req-bbc50369-5499-4d5a-bfdd-dad20fe3064d service nova] Acquiring lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.143979] env[61906]: DEBUG oslo_concurrency.lockutils [req-c8394cfe-37f6-4002-a002-5ab639e7318d req-bbc50369-5499-4d5a-bfdd-dad20fe3064d service nova] Lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.143979] env[61906]: DEBUG oslo_concurrency.lockutils [req-c8394cfe-37f6-4002-a002-5ab639e7318d req-bbc50369-5499-4d5a-bfdd-dad20fe3064d service nova] Lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.143979] env[61906]: DEBUG nova.compute.manager [req-c8394cfe-37f6-4002-a002-5ab639e7318d req-bbc50369-5499-4d5a-bfdd-dad20fe3064d service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] No waiting events found dispatching network-vif-plugged-c2232d91-1c2c-4a66-9026-933169da90e1 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 929.145038] env[61906]: WARNING nova.compute.manager [req-c8394cfe-37f6-4002-a002-5ab639e7318d req-bbc50369-5499-4d5a-bfdd-dad20fe3064d service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Received unexpected event network-vif-plugged-c2232d91-1c2c-4a66-9026-933169da90e1 for instance with vm_state active and task_state None. [ 929.372007] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333374, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071134} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.372323] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 929.373517] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ea9972c-6ce2-427a-a7bc-ee98305dcb37 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.395784] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 11cb9d89-2304-45a6-b4bb-fcf8417a0518/11cb9d89-2304-45a6-b4bb-fcf8417a0518.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 929.396135] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ce75604-3afa-43ec-86a7-12fa8ccc5a1f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.417016] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 929.417016] env[61906]: value = "task-1333376" [ 929.417016] env[61906]: _type = "Task" [ 929.417016] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.428242] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333376, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.507354] env[61906]: DEBUG oslo_concurrency.lockutils [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.800s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.510273] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 11.213s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.534881] env[61906]: DEBUG oslo_vmware.api [None req-54cca8b8-a2a9-4212-b03b-fd79a7202db3 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333375, 'name': SuspendVM_Task} progress is 58%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.542366] env[61906]: INFO nova.scheduler.client.report [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Deleted allocations for instance 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72 [ 929.555077] env[61906]: DEBUG oslo_vmware.api [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333371, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.382003} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.555869] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 929.556705] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 929.556936] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 929.557171] env[61906]: INFO nova.compute.manager [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Took 2.15 seconds to destroy the instance on the hypervisor. [ 929.557471] env[61906]: DEBUG oslo.service.loopingcall [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 929.557732] env[61906]: DEBUG nova.compute.manager [-] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 929.557853] env[61906]: DEBUG nova.network.neutron [-] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 929.604195] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]525dbd61-2d2e-98ee-4f38-8c6cdd86ecfb, 'name': SearchDatastore_Task, 'duration_secs': 0.015975} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.604748] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8cc02bd0-ec1d-4efe-b1b1-56c51aad5be9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.612029] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 929.612029] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52326531-9811-cbc1-8bfe-e66944cc92a5" [ 929.612029] env[61906]: _type = "Task" [ 929.612029] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.630024] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52326531-9811-cbc1-8bfe-e66944cc92a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.633096] env[61906]: INFO nova.compute.manager [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Took 15.99 seconds to build instance. [ 929.634303] env[61906]: DEBUG oslo_concurrency.lockutils [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "refresh_cache-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.634479] env[61906]: DEBUG oslo_concurrency.lockutils [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired lock "refresh_cache-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.634654] env[61906]: DEBUG nova.network.neutron [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 929.932148] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333376, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.021150] env[61906]: INFO nova.compute.claims [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 930.040980] env[61906]: DEBUG oslo_vmware.api [None req-54cca8b8-a2a9-4212-b03b-fd79a7202db3 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333375, 'name': SuspendVM_Task, 'duration_secs': 0.726885} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.041259] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-54cca8b8-a2a9-4212-b03b-fd79a7202db3 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Suspended the VM {{(pid=61906) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 930.045091] env[61906]: DEBUG nova.compute.manager [None req-54cca8b8-a2a9-4212-b03b-fd79a7202db3 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 930.047226] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95f0d0c0-73a0-473b-b2ea-63aab54b1257 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.062259] env[61906]: DEBUG oslo_concurrency.lockutils [None req-39575535-7cc3-4085-b93b-7f73222a4ddb tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "1bc9f3d4-b539-4841-b5b6-eaefa0d06a72" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.497s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.122499] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52326531-9811-cbc1-8bfe-e66944cc92a5, 'name': SearchDatastore_Task, 'duration_secs': 0.011577} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.122847] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.123144] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5/8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 930.127020] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1a9f3c5e-aee6-4e21-a3a4-15c26b65cbee {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.131340] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 930.131340] env[61906]: value = "task-1333377" [ 930.131340] env[61906]: _type = "Task" [ 930.131340] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.135156] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eff62f6d-4362-41bd-9ff4-8d7c09b855f0 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "1a335019-915b-4065-b22f-d2f0af728d6f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.511s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.142109] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333377, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.204844] env[61906]: WARNING nova.network.neutron [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] 37d8737f-db44-4cf6-8e34-31b4cbd2f82e already exists in list: networks containing: ['37d8737f-db44-4cf6-8e34-31b4cbd2f82e']. ignoring it [ 930.350147] env[61906]: DEBUG nova.network.neutron [-] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.428376] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333376, 'name': ReconfigVM_Task, 'duration_secs': 0.860881} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.428851] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 11cb9d89-2304-45a6-b4bb-fcf8417a0518/11cb9d89-2304-45a6-b4bb-fcf8417a0518.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 930.429570] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e49ea051-01f2-457f-976a-87ec9df2ffa5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.439081] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 930.439081] env[61906]: value = "task-1333378" [ 930.439081] env[61906]: _type = "Task" [ 930.439081] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.452074] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333378, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.530912] env[61906]: INFO nova.compute.resource_tracker [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Updating resource usage from migration f037f355-83fb-4668-9837-81e957b39c1e [ 930.644876] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333377, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.482144} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.645235] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5/8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 930.645721] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 930.645721] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3fdd84ce-3a1e-4aa9-b8f0-6cb50196ecd8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.655837] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 930.655837] env[61906]: value = "task-1333379" [ 930.655837] env[61906]: _type = "Task" [ 930.655837] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.666182] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333379, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.757452] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3df63641-2046-4a29-b733-7dd132a56c01 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.765936] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-000d4f86-db21-4e1b-a636-cf7a815cb418 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.798502] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c2ac489-9a5d-4ec5-8ba5-142e41ccca60 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.810024] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08ffaea0-9bd2-4e9f-bdcb-4c99d4151127 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.823571] env[61906]: DEBUG nova.compute.provider_tree [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.856500] env[61906]: INFO nova.compute.manager [-] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Took 1.30 seconds to deallocate network for instance. [ 930.949578] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333378, 'name': Rename_Task, 'duration_secs': 0.201376} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.949896] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 930.950161] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3d2b83b3-6d17-4dc9-b6af-28442a9a5273 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.957761] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 930.957761] env[61906]: value = "task-1333380" [ 930.957761] env[61906]: _type = "Task" [ 930.957761] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.966654] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333380, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.009182] env[61906]: DEBUG nova.network.neutron [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Updating instance_info_cache with network_info: [{"id": "3e04c5c7-8d7a-46ae-a2d7-52dc429650f9", "address": "fa:16:3e:f1:15:76", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e04c5c7-8d", "ovs_interfaceid": "3e04c5c7-8d7a-46ae-a2d7-52dc429650f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c2232d91-1c2c-4a66-9026-933169da90e1", "address": "fa:16:3e:80:0d:49", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2232d91-1c", "ovs_interfaceid": "c2232d91-1c2c-4a66-9026-933169da90e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.168463] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333379, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07349} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.168463] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 931.168463] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba3340c-e6d8-4378-806d-c849f6986d3e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.188070] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5/8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 931.188413] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-588f976b-5747-412d-a037-6e902fd6dd8f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.209533] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 931.209533] env[61906]: value = "task-1333381" [ 931.209533] env[61906]: _type = "Task" [ 931.209533] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.219433] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333381, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.327308] env[61906]: DEBUG nova.scheduler.client.report [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 931.363124] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.469699] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333380, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.513122] env[61906]: DEBUG oslo_concurrency.lockutils [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Releasing lock "refresh_cache-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.513852] env[61906]: DEBUG oslo_concurrency.lockutils [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.513993] env[61906]: DEBUG oslo_concurrency.lockutils [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.514936] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35bcfd6c-f679-4c31-92fb-22710ab3d3a7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.543544] env[61906]: DEBUG nova.virt.hardware [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 931.544110] env[61906]: DEBUG nova.virt.hardware [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 931.544317] env[61906]: DEBUG nova.virt.hardware [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 931.547101] env[61906]: DEBUG nova.virt.hardware [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 931.547101] env[61906]: DEBUG nova.virt.hardware [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 931.547101] env[61906]: DEBUG nova.virt.hardware [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 931.547101] env[61906]: DEBUG nova.virt.hardware [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 931.547101] env[61906]: DEBUG nova.virt.hardware [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 931.547101] env[61906]: DEBUG nova.virt.hardware [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 931.547101] env[61906]: DEBUG nova.virt.hardware [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 931.547101] env[61906]: DEBUG nova.virt.hardware [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 931.558936] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Reconfiguring VM to attach interface {{(pid=61906) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 931.558936] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-08332368-48b8-4deb-b7f4-d303bb3b1475 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.579143] env[61906]: DEBUG oslo_vmware.api [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 931.579143] env[61906]: value = "task-1333382" [ 931.579143] env[61906]: _type = "Task" [ 931.579143] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.592586] env[61906]: DEBUG oslo_vmware.api [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333382, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.681547] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "83b59194-b86f-416a-a7fd-18f75551c2c0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.681814] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "83b59194-b86f-416a-a7fd-18f75551c2c0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.721988] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333381, 'name': ReconfigVM_Task, 'duration_secs': 0.388255} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.722598] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5/8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 931.723424] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-07c3d169-de47-44e8-9413-c87727416110 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.732523] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 931.732523] env[61906]: value = "task-1333383" [ 931.732523] env[61906]: _type = "Task" [ 931.732523] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.741447] env[61906]: DEBUG nova.compute.manager [req-0127ee4c-5257-4005-b1ff-222cd2b28bb8 req-d9b5dc8f-3a65-4123-b387-330ab32719bf service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Received event network-changed-c2232d91-1c2c-4a66-9026-933169da90e1 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 931.741447] env[61906]: DEBUG nova.compute.manager [req-0127ee4c-5257-4005-b1ff-222cd2b28bb8 req-d9b5dc8f-3a65-4123-b387-330ab32719bf service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Refreshing instance network info cache due to event network-changed-c2232d91-1c2c-4a66-9026-933169da90e1. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 931.741447] env[61906]: DEBUG oslo_concurrency.lockutils [req-0127ee4c-5257-4005-b1ff-222cd2b28bb8 req-d9b5dc8f-3a65-4123-b387-330ab32719bf service nova] Acquiring lock "refresh_cache-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.741447] env[61906]: DEBUG oslo_concurrency.lockutils [req-0127ee4c-5257-4005-b1ff-222cd2b28bb8 req-d9b5dc8f-3a65-4123-b387-330ab32719bf service nova] Acquired lock "refresh_cache-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.741447] env[61906]: DEBUG nova.network.neutron [req-0127ee4c-5257-4005-b1ff-222cd2b28bb8 req-d9b5dc8f-3a65-4123-b387-330ab32719bf service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Refreshing network info cache for port c2232d91-1c2c-4a66-9026-933169da90e1 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 931.745811] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333383, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.832963] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.323s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.833209] env[61906]: INFO nova.compute.manager [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Migrating [ 931.833468] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.833621] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired lock "compute-rpcapi-router" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.835617] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.485s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.835617] env[61906]: DEBUG nova.objects.instance [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Lazy-loading 'resources' on Instance uuid d60575a6-af13-4d81-95d1-081d0f3e2ce6 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 931.969212] env[61906]: DEBUG oslo_vmware.api [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333380, 'name': PowerOnVM_Task, 'duration_secs': 0.890853} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.969504] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 931.969712] env[61906]: INFO nova.compute.manager [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Took 6.34 seconds to spawn the instance on the hypervisor. [ 931.969923] env[61906]: DEBUG nova.compute.manager [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 931.970818] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ab875a2-3ecf-4892-9aae-3d7183d2377e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.089710] env[61906]: DEBUG oslo_vmware.api [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333382, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.184559] env[61906]: DEBUG nova.compute.manager [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 932.245359] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333383, 'name': Rename_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.337135] env[61906]: INFO nova.compute.rpcapi [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 932.337925] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Releasing lock "compute-rpcapi-router" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.495594] env[61906]: INFO nova.compute.manager [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Took 17.50 seconds to build instance. [ 932.556294] env[61906]: DEBUG nova.network.neutron [req-0127ee4c-5257-4005-b1ff-222cd2b28bb8 req-d9b5dc8f-3a65-4123-b387-330ab32719bf service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Updated VIF entry in instance network info cache for port c2232d91-1c2c-4a66-9026-933169da90e1. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 932.557085] env[61906]: DEBUG nova.network.neutron [req-0127ee4c-5257-4005-b1ff-222cd2b28bb8 req-d9b5dc8f-3a65-4123-b387-330ab32719bf service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Updating instance_info_cache with network_info: [{"id": "3e04c5c7-8d7a-46ae-a2d7-52dc429650f9", "address": "fa:16:3e:f1:15:76", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e04c5c7-8d", "ovs_interfaceid": "3e04c5c7-8d7a-46ae-a2d7-52dc429650f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c2232d91-1c2c-4a66-9026-933169da90e1", "address": "fa:16:3e:80:0d:49", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2232d91-1c", "ovs_interfaceid": "c2232d91-1c2c-4a66-9026-933169da90e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.593185] env[61906]: DEBUG oslo_vmware.api [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333382, 'name': ReconfigVM_Task, 'duration_secs': 0.680998} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.593715] env[61906]: DEBUG oslo_concurrency.lockutils [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Releasing lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.593951] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Reconfigured VM to attach interface {{(pid=61906) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 932.620851] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1947f61-169e-409d-a35c-2014973a9a57 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.631134] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d4b8cf1-fd02-47fa-90e5-177e1d5bbed4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.665808] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a84e22a-bd4b-4107-bd8f-fbe1f44066d5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.674917] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cb5b7e6-a1ed-4413-858e-b6fb6a116b6e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.697187] env[61906]: DEBUG nova.compute.provider_tree [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 932.709394] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.743180] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333383, 'name': Rename_Task, 'duration_secs': 0.972186} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.743819] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 932.743819] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-197d4674-2dbb-4f29-b36a-306278e1946c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.751546] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 932.751546] env[61906]: value = "task-1333384" [ 932.751546] env[61906]: _type = "Task" [ 932.751546] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.760426] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333384, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.867268] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "refresh_cache-e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.867480] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired lock "refresh_cache-e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.867666] env[61906]: DEBUG nova.network.neutron [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 932.922852] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Acquiring lock "da493512-d996-4de7-9f47-cadcbc4fbcb9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.923657] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Lock "da493512-d996-4de7-9f47-cadcbc4fbcb9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.923657] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Acquiring lock "da493512-d996-4de7-9f47-cadcbc4fbcb9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.923657] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Lock "da493512-d996-4de7-9f47-cadcbc4fbcb9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.924845] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Lock "da493512-d996-4de7-9f47-cadcbc4fbcb9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.926202] env[61906]: INFO nova.compute.manager [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Terminating instance [ 932.928069] env[61906]: DEBUG nova.compute.manager [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 932.928305] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 932.928570] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d5903eff-5c59-41dd-a5d2-e5e367cf03af {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.936881] env[61906]: DEBUG oslo_vmware.api [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Waiting for the task: (returnval){ [ 932.936881] env[61906]: value = "task-1333385" [ 932.936881] env[61906]: _type = "Task" [ 932.936881] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.945664] env[61906]: DEBUG oslo_vmware.api [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Task: {'id': task-1333385, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.997067] env[61906]: DEBUG oslo_concurrency.lockutils [None req-420dc29c-e30c-4728-9749-213e27e6c827 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Lock "11cb9d89-2304-45a6-b4bb-fcf8417a0518" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.009s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.066145] env[61906]: DEBUG oslo_concurrency.lockutils [req-0127ee4c-5257-4005-b1ff-222cd2b28bb8 req-d9b5dc8f-3a65-4123-b387-330ab32719bf service nova] Releasing lock "refresh_cache-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.067416] env[61906]: DEBUG nova.compute.manager [req-0127ee4c-5257-4005-b1ff-222cd2b28bb8 req-d9b5dc8f-3a65-4123-b387-330ab32719bf service nova] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Received event network-vif-deleted-b0492e76-fbe6-486e-8fe8-539b1e22ab50 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 933.067416] env[61906]: DEBUG nova.compute.manager [req-0127ee4c-5257-4005-b1ff-222cd2b28bb8 req-d9b5dc8f-3a65-4123-b387-330ab32719bf service nova] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Received event network-changed-a497b774-60ea-485d-a564-ba7b978b560e {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 933.067416] env[61906]: DEBUG nova.compute.manager [req-0127ee4c-5257-4005-b1ff-222cd2b28bb8 req-d9b5dc8f-3a65-4123-b387-330ab32719bf service nova] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Refreshing instance network info cache due to event network-changed-a497b774-60ea-485d-a564-ba7b978b560e. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 933.067416] env[61906]: DEBUG oslo_concurrency.lockutils [req-0127ee4c-5257-4005-b1ff-222cd2b28bb8 req-d9b5dc8f-3a65-4123-b387-330ab32719bf service nova] Acquiring lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.067416] env[61906]: DEBUG oslo_concurrency.lockutils [req-0127ee4c-5257-4005-b1ff-222cd2b28bb8 req-d9b5dc8f-3a65-4123-b387-330ab32719bf service nova] Acquired lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.067416] env[61906]: DEBUG nova.network.neutron [req-0127ee4c-5257-4005-b1ff-222cd2b28bb8 req-d9b5dc8f-3a65-4123-b387-330ab32719bf service nova] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Refreshing network info cache for port a497b774-60ea-485d-a564-ba7b978b560e {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 933.099162] env[61906]: DEBUG oslo_concurrency.lockutils [None req-631904f7-6f8e-42d0-ad9f-88c66cfcab24 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "interface-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.898s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.201559] env[61906]: DEBUG nova.scheduler.client.report [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 933.262687] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333384, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.300593] env[61906]: DEBUG nova.compute.manager [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 933.301653] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6218a347-5843-470f-ab7d-48b057702ba2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.446776] env[61906]: DEBUG oslo_vmware.api [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Task: {'id': task-1333385, 'name': PowerOffVM_Task, 'duration_secs': 0.218309} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.449298] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 933.449551] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Volume detach. Driver type: vmdk {{(pid=61906) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 933.449840] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284737', 'volume_id': '7f04fa3d-ec3b-4151-a46b-f232180344b9', 'name': 'volume-7f04fa3d-ec3b-4151-a46b-f232180344b9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'da493512-d996-4de7-9f47-cadcbc4fbcb9', 'attached_at': '', 'detached_at': '', 'volume_id': '7f04fa3d-ec3b-4151-a46b-f232180344b9', 'serial': '7f04fa3d-ec3b-4151-a46b-f232180344b9'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 933.450709] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43709871-f12f-4165-9fdd-df956bceeabb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.474190] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36798a20-389f-4535-a67a-f7a7779c444a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.481473] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff5accf8-4441-452d-a241-fc0a4ca5202e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.503075] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fc89e6c-517d-4da8-a6b2-e441bb3d9b13 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.519798] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] The volume has not been displaced from its original location: [datastore2] volume-7f04fa3d-ec3b-4151-a46b-f232180344b9/volume-7f04fa3d-ec3b-4151-a46b-f232180344b9.vmdk. No consolidation needed. {{(pid=61906) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 933.525564] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Reconfiguring VM instance instance-00000047 to detach disk 2000 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 933.525951] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82f9ecfe-20d7-4b7c-97af-a17c640e39e2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.549466] env[61906]: DEBUG oslo_vmware.api [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Waiting for the task: (returnval){ [ 933.549466] env[61906]: value = "task-1333386" [ 933.549466] env[61906]: _type = "Task" [ 933.549466] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.558262] env[61906]: DEBUG oslo_vmware.api [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Task: {'id': task-1333386, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.618632] env[61906]: DEBUG nova.network.neutron [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Updating instance_info_cache with network_info: [{"id": "b54c27f8-7727-4166-af82-5b6fe0d3519e", "address": "fa:16:3e:ae:b6:e5", "network": {"id": "af039bdd-ffd2-483d-9655-e7df48936290", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-111531540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f7fc2042e6549cabe98838990f35842", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb54c27f8-77", "ovs_interfaceid": "b54c27f8-7727-4166-af82-5b6fe0d3519e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.695614] env[61906]: DEBUG oslo_concurrency.lockutils [None req-999a1791-dfd9-4a13-9a0d-d223f011b2c5 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "7504dfa6-bedc-4701-b4fc-60e19e742276" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.695864] env[61906]: DEBUG oslo_concurrency.lockutils [None req-999a1791-dfd9-4a13-9a0d-d223f011b2c5 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "7504dfa6-bedc-4701-b4fc-60e19e742276" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.710378] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.875s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.712906] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 6.174s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.712906] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.713454] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61906) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 933.713454] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.350s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.713576] env[61906]: DEBUG nova.objects.instance [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lazy-loading 'resources' on Instance uuid 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 933.715562] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffcc3930-640c-4657-b551-ba3761584ebe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.725784] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c25c39c-fcf6-4feb-98c3-c0be3bd53175 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.743907] env[61906]: INFO nova.scheduler.client.report [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Deleted allocations for instance d60575a6-af13-4d81-95d1-081d0f3e2ce6 [ 933.745676] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3ceb836-12d0-4a9a-99a7-a7d2f5eb46b7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.758051] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c961730-ca5f-4df1-98ba-1d75d3997734 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.766703] env[61906]: DEBUG oslo_vmware.api [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333384, 'name': PowerOnVM_Task, 'duration_secs': 0.926502} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.792794] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 933.793068] env[61906]: INFO nova.compute.manager [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Took 5.84 seconds to spawn the instance on the hypervisor. [ 933.793353] env[61906]: DEBUG nova.compute.manager [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 933.794336] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181428MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=61906) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 933.794479] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.797510] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05df69f6-d7ed-4f57-98fb-133cce745067 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.813601] env[61906]: INFO nova.compute.manager [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] instance snapshotting [ 933.813601] env[61906]: WARNING nova.compute.manager [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 933.816437] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc1d6e8f-288b-4c88-a07d-886dd2c9069b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.835971] env[61906]: DEBUG nova.network.neutron [req-0127ee4c-5257-4005-b1ff-222cd2b28bb8 req-d9b5dc8f-3a65-4123-b387-330ab32719bf service nova] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Updated VIF entry in instance network info cache for port a497b774-60ea-485d-a564-ba7b978b560e. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 933.836348] env[61906]: DEBUG nova.network.neutron [req-0127ee4c-5257-4005-b1ff-222cd2b28bb8 req-d9b5dc8f-3a65-4123-b387-330ab32719bf service nova] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Updating instance_info_cache with network_info: [{"id": "a497b774-60ea-485d-a564-ba7b978b560e", "address": "fa:16:3e:0d:cf:6e", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa497b774-60", "ovs_interfaceid": "a497b774-60ea-485d-a564-ba7b978b560e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.837893] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b6fe147-4600-41ce-aa4b-75f216784d86 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.060326] env[61906]: DEBUG oslo_vmware.api [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Task: {'id': task-1333386, 'name': ReconfigVM_Task, 'duration_secs': 0.236779} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.060528] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Reconfigured VM instance instance-00000047 to detach disk 2000 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 934.065431] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4e988cfe-0e89-4e41-8d98-0c2326a98ff0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.082420] env[61906]: DEBUG oslo_vmware.api [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Waiting for the task: (returnval){ [ 934.082420] env[61906]: value = "task-1333387" [ 934.082420] env[61906]: _type = "Task" [ 934.082420] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.096500] env[61906]: DEBUG oslo_vmware.api [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Task: {'id': task-1333387, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.121589] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Releasing lock "refresh_cache-e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.199390] env[61906]: DEBUG nova.compute.utils [None req-999a1791-dfd9-4a13-9a0d-d223f011b2c5 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 934.256854] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ad708f7-8659-4177-87d4-9517d0ec001f tempest-ImagesOneServerTestJSON-1620714537 tempest-ImagesOneServerTestJSON-1620714537-project-member] Lock "d60575a6-af13-4d81-95d1-081d0f3e2ce6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.859s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.321134] env[61906]: INFO nova.compute.manager [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Took 19.06 seconds to build instance. [ 934.343638] env[61906]: DEBUG oslo_concurrency.lockutils [req-0127ee4c-5257-4005-b1ff-222cd2b28bb8 req-d9b5dc8f-3a65-4123-b387-330ab32719bf service nova] Releasing lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.349588] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Creating Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 934.349968] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-cc50fe14-8971-4a46-adcc-297516473f9d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.366695] env[61906]: DEBUG oslo_vmware.api [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 934.366695] env[61906]: value = "task-1333388" [ 934.366695] env[61906]: _type = "Task" [ 934.366695] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.377679] env[61906]: DEBUG oslo_vmware.api [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333388, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.458218] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3945a82d-f2c8-4b14-a514-530ce73339d6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.467122] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a2719b2-5ee3-406c-b6ba-8539ecafba8e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.498812] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-341ab616-982a-42b5-a150-e94d0e45b29a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.508274] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b76a0f2-1061-440c-a993-ebecc7afff0f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.523690] env[61906]: DEBUG nova.compute.provider_tree [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 934.593624] env[61906]: DEBUG oslo_vmware.api [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Task: {'id': task-1333387, 'name': ReconfigVM_Task, 'duration_secs': 0.159804} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.594033] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284737', 'volume_id': '7f04fa3d-ec3b-4151-a46b-f232180344b9', 'name': 'volume-7f04fa3d-ec3b-4151-a46b-f232180344b9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'da493512-d996-4de7-9f47-cadcbc4fbcb9', 'attached_at': '', 'detached_at': '', 'volume_id': '7f04fa3d-ec3b-4151-a46b-f232180344b9', 'serial': '7f04fa3d-ec3b-4151-a46b-f232180344b9'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 934.594392] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 934.595318] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9680be5f-7bf8-45ff-ab68-80f7f45f8496 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.603818] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 934.604167] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b1659878-b32b-4abd-9cf8-8716bbd26826 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.668657] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 934.668846] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 934.669018] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Deleting the datastore file [datastore2] da493512-d996-4de7-9f47-cadcbc4fbcb9 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 934.669318] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-12aed908-30c6-4b8c-9ee6-67304e2d3dd9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.679491] env[61906]: DEBUG oslo_vmware.api [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Waiting for the task: (returnval){ [ 934.679491] env[61906]: value = "task-1333390" [ 934.679491] env[61906]: _type = "Task" [ 934.679491] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.688644] env[61906]: DEBUG oslo_vmware.api [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Task: {'id': task-1333390, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.703479] env[61906]: DEBUG oslo_concurrency.lockutils [None req-999a1791-dfd9-4a13-9a0d-d223f011b2c5 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "7504dfa6-bedc-4701-b4fc-60e19e742276" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.824235] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc2284c1-9a47-4328-a019-ac96fcd9bd54 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Lock "8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.570s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.879735] env[61906]: DEBUG oslo_vmware.api [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333388, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.938281] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "interface-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8-4f8ab1c1-fd91-48d3-90e0-ed5ff9fa0b4d" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.938281] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "interface-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8-4f8ab1c1-fd91-48d3-90e0-ed5ff9fa0b4d" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.938476] env[61906]: DEBUG nova.objects.instance [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lazy-loading 'flavor' on Instance uuid a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.027577] env[61906]: DEBUG nova.scheduler.client.report [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 935.190407] env[61906]: DEBUG oslo_vmware.api [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Task: {'id': task-1333390, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.092875} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.190706] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.190902] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 935.191130] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 935.191278] env[61906]: INFO nova.compute.manager [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Took 2.26 seconds to destroy the instance on the hypervisor. [ 935.191528] env[61906]: DEBUG oslo.service.loopingcall [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 935.191852] env[61906]: DEBUG nova.compute.manager [-] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 935.191852] env[61906]: DEBUG nova.network.neutron [-] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 935.379331] env[61906]: DEBUG oslo_vmware.api [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333388, 'name': CreateSnapshot_Task, 'duration_secs': 0.648978} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.379658] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Created Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 935.380715] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d186026-bf2e-4c75-bcbd-68a6a8eb703a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.533060] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.819s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.537330] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.826s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.537803] env[61906]: INFO nova.compute.claims [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 935.566260] env[61906]: INFO nova.scheduler.client.report [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Deleted allocations for instance 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a [ 935.625299] env[61906]: DEBUG nova.objects.instance [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lazy-loading 'pci_requests' on Instance uuid a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.642306] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e530b30-97ae-47bf-b584-4a449761359c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.664438] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Updating instance 'e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15' progress to 0 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 935.784595] env[61906]: DEBUG oslo_concurrency.lockutils [None req-999a1791-dfd9-4a13-9a0d-d223f011b2c5 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "7504dfa6-bedc-4701-b4fc-60e19e742276" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.784595] env[61906]: DEBUG oslo_concurrency.lockutils [None req-999a1791-dfd9-4a13-9a0d-d223f011b2c5 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "7504dfa6-bedc-4701-b4fc-60e19e742276" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.784595] env[61906]: INFO nova.compute.manager [None req-999a1791-dfd9-4a13-9a0d-d223f011b2c5 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Attaching volume a8641678-cc5c-458c-a0a6-daf5317bf7d7 to /dev/sdb [ 935.823723] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61482785-0511-4153-a36b-51ac17cd6631 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.835049] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b20b861-c445-4bfc-83c9-b6a48f302b9d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.851556] env[61906]: DEBUG nova.virt.block_device [None req-999a1791-dfd9-4a13-9a0d-d223f011b2c5 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Updating existing volume attachment record: 3bec54ba-46ea-4ee5-9c00-a0154e7fa1cc {{(pid=61906) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 935.904835] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Creating linked-clone VM from snapshot {{(pid=61906) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 935.906974] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d9f32d09-be14-4b60-a9f7-7db88b4ea238 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.916570] env[61906]: DEBUG oslo_vmware.api [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 935.916570] env[61906]: value = "task-1333391" [ 935.916570] env[61906]: _type = "Task" [ 935.916570] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.926365] env[61906]: DEBUG oslo_vmware.api [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333391, 'name': CloneVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.006329] env[61906]: DEBUG nova.compute.manager [req-1763b782-c588-4f33-a1c4-007e8f73c98d req-c6e0c611-a654-474e-9c59-f85c6147b2fa service nova] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Received event network-vif-deleted-f76ea64e-15c4-490e-9a5e-4c244195d983 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 936.006510] env[61906]: INFO nova.compute.manager [req-1763b782-c588-4f33-a1c4-007e8f73c98d req-c6e0c611-a654-474e-9c59-f85c6147b2fa service nova] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Neutron deleted interface f76ea64e-15c4-490e-9a5e-4c244195d983; detaching it from the instance and deleting it from the info cache [ 936.006876] env[61906]: DEBUG nova.network.neutron [req-1763b782-c588-4f33-a1c4-007e8f73c98d req-c6e0c611-a654-474e-9c59-f85c6147b2fa service nova] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.081797] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8d12fdab-0edb-4e07-bf93-ae9aaffeb5ee tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "0488c6eb-b510-4d33-ab9f-ebfdab63fb3a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.679s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.129992] env[61906]: DEBUG nova.objects.base [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 936.130471] env[61906]: DEBUG nova.network.neutron [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 936.174651] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 936.175140] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2a3a2ddd-419a-430b-9d68-0df836a6fc0c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.186690] env[61906]: DEBUG oslo_vmware.api [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 936.186690] env[61906]: value = "task-1333395" [ 936.186690] env[61906]: _type = "Task" [ 936.186690] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.198960] env[61906]: DEBUG oslo_vmware.api [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333395, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.225613] env[61906]: DEBUG nova.policy [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0de9abfd085f4a43a878da8448615bc4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e36ff7a500ef444cbf5a168c5a48208b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 936.410800] env[61906]: DEBUG nova.network.neutron [-] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.431748] env[61906]: DEBUG oslo_vmware.api [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333391, 'name': CloneVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.499104] env[61906]: INFO nova.compute.manager [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Rebuilding instance [ 936.509898] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-974b1605-f489-47f0-98b4-b907df6d7980 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.525656] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e97c34d-eeac-4dea-8380-1dc2b3082cf8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.568186] env[61906]: DEBUG nova.compute.manager [req-1763b782-c588-4f33-a1c4-007e8f73c98d req-c6e0c611-a654-474e-9c59-f85c6147b2fa service nova] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Detach interface failed, port_id=f76ea64e-15c4-490e-9a5e-4c244195d983, reason: Instance da493512-d996-4de7-9f47-cadcbc4fbcb9 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 936.577372] env[61906]: DEBUG nova.compute.manager [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 936.578601] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97b36324-19bf-481d-abcb-f26358bf6640 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.697390] env[61906]: DEBUG oslo_vmware.api [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333395, 'name': PowerOffVM_Task, 'duration_secs': 0.22794} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.700345] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 936.700437] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Updating instance 'e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15' progress to 17 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 936.729101] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98589165-5678-4d7e-b530-522d54e5a00f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.737816] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d94fb6a-819b-415c-a7d8-7a3792b1e9b4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.771157] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7e5bf7a-dc80-496b-85f3-814ef0d609f4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.782537] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78c61c06-ea21-41c0-bfda-dc4e4a722814 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.799394] env[61906]: DEBUG nova.compute.provider_tree [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.916632] env[61906]: INFO nova.compute.manager [-] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Took 1.72 seconds to deallocate network for instance. [ 936.928844] env[61906]: DEBUG oslo_vmware.api [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333391, 'name': CloneVM_Task} progress is 95%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.093970] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 937.094325] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b8f16fbf-7a8f-4b59-b1a2-441343ba03e4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.102143] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 937.102143] env[61906]: value = "task-1333396" [ 937.102143] env[61906]: _type = "Task" [ 937.102143] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.113955] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333396, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.208119] env[61906]: DEBUG nova.virt.hardware [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:23Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 937.208431] env[61906]: DEBUG nova.virt.hardware [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 937.208609] env[61906]: DEBUG nova.virt.hardware [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 937.208793] env[61906]: DEBUG nova.virt.hardware [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 937.208933] env[61906]: DEBUG nova.virt.hardware [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 937.209095] env[61906]: DEBUG nova.virt.hardware [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 937.209302] env[61906]: DEBUG nova.virt.hardware [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 937.209461] env[61906]: DEBUG nova.virt.hardware [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 937.212787] env[61906]: DEBUG nova.virt.hardware [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 937.213498] env[61906]: DEBUG nova.virt.hardware [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 937.213498] env[61906]: DEBUG nova.virt.hardware [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 937.219373] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4376de48-7903-4efe-924b-51d288eadfda {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.240466] env[61906]: DEBUG oslo_vmware.api [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 937.240466] env[61906]: value = "task-1333397" [ 937.240466] env[61906]: _type = "Task" [ 937.240466] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.979265] env[61906]: DEBUG nova.scheduler.client.report [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 937.983139] env[61906]: DEBUG oslo_vmware.api [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333397, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.983139] env[61906]: WARNING oslo_vmware.common.loopingcall [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] task run outlasted interval by 0.243008 sec [ 937.983813] env[61906]: DEBUG nova.network.neutron [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Successfully updated port: 4f8ab1c1-fd91-48d3-90e0-ed5ff9fa0b4d {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 937.986445] env[61906]: INFO nova.compute.manager [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Took 1.07 seconds to detach 1 volumes for instance. [ 937.996520] env[61906]: DEBUG nova.compute.manager [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Deleting volume: 7f04fa3d-ec3b-4151-a46b-f232180344b9 {{(pid=61906) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 938.002109] env[61906]: DEBUG oslo_concurrency.lockutils [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "0f846f95-2573-4450-9918-a34467d73363" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.002109] env[61906]: DEBUG oslo_concurrency.lockutils [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "0f846f95-2573-4450-9918-a34467d73363" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.016917] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333396, 'name': PowerOffVM_Task, 'duration_secs': 0.141983} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.017839] env[61906]: DEBUG oslo_vmware.api [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333397, 'name': ReconfigVM_Task, 'duration_secs': 0.343045} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.018323] env[61906]: DEBUG oslo_vmware.api [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333391, 'name': CloneVM_Task, 'duration_secs': 1.436591} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.018552] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 938.018764] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 938.019093] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Updating instance 'e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15' progress to 33 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 938.022251] env[61906]: INFO nova.virt.vmwareapi.vmops [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Created linked-clone VM from snapshot [ 938.023287] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76608569-d66a-4cb5-aa52-a879fd31ff3e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.028068] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a864712-941c-42f6-b3d3-e0318fb42421 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.037070] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Uploading image 670311c4-3438-463b-9026-04bf6b5eb856 {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 938.043192] env[61906]: DEBUG nova.compute.manager [req-47754955-eb75-4e5c-8bd7-a7669509fd8b req-ab1f0894-3d33-4d09-a65d-347c4c0b3699 service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Received event network-vif-plugged-4f8ab1c1-fd91-48d3-90e0-ed5ff9fa0b4d {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 938.043192] env[61906]: DEBUG oslo_concurrency.lockutils [req-47754955-eb75-4e5c-8bd7-a7669509fd8b req-ab1f0894-3d33-4d09-a65d-347c4c0b3699 service nova] Acquiring lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.043192] env[61906]: DEBUG oslo_concurrency.lockutils [req-47754955-eb75-4e5c-8bd7-a7669509fd8b req-ab1f0894-3d33-4d09-a65d-347c4c0b3699 service nova] Lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.043192] env[61906]: DEBUG oslo_concurrency.lockutils [req-47754955-eb75-4e5c-8bd7-a7669509fd8b req-ab1f0894-3d33-4d09-a65d-347c4c0b3699 service nova] Lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.043192] env[61906]: DEBUG nova.compute.manager [req-47754955-eb75-4e5c-8bd7-a7669509fd8b req-ab1f0894-3d33-4d09-a65d-347c4c0b3699 service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] No waiting events found dispatching network-vif-plugged-4f8ab1c1-fd91-48d3-90e0-ed5ff9fa0b4d {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 938.043192] env[61906]: WARNING nova.compute.manager [req-47754955-eb75-4e5c-8bd7-a7669509fd8b req-ab1f0894-3d33-4d09-a65d-347c4c0b3699 service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Received unexpected event network-vif-plugged-4f8ab1c1-fd91-48d3-90e0-ed5ff9fa0b4d for instance with vm_state active and task_state None. [ 938.043617] env[61906]: DEBUG nova.compute.manager [req-47754955-eb75-4e5c-8bd7-a7669509fd8b req-ab1f0894-3d33-4d09-a65d-347c4c0b3699 service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Received event network-changed-4f8ab1c1-fd91-48d3-90e0-ed5ff9fa0b4d {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 938.043617] env[61906]: DEBUG nova.compute.manager [req-47754955-eb75-4e5c-8bd7-a7669509fd8b req-ab1f0894-3d33-4d09-a65d-347c4c0b3699 service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Refreshing instance network info cache due to event network-changed-4f8ab1c1-fd91-48d3-90e0-ed5ff9fa0b4d. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 938.043617] env[61906]: DEBUG oslo_concurrency.lockutils [req-47754955-eb75-4e5c-8bd7-a7669509fd8b req-ab1f0894-3d33-4d09-a65d-347c4c0b3699 service nova] Acquiring lock "refresh_cache-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.043816] env[61906]: DEBUG oslo_concurrency.lockutils [req-47754955-eb75-4e5c-8bd7-a7669509fd8b req-ab1f0894-3d33-4d09-a65d-347c4c0b3699 service nova] Acquired lock "refresh_cache-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.043973] env[61906]: DEBUG nova.network.neutron [req-47754955-eb75-4e5c-8bd7-a7669509fd8b req-ab1f0894-3d33-4d09-a65d-347c4c0b3699 service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Refreshing network info cache for port 4f8ab1c1-fd91-48d3-90e0-ed5ff9fa0b4d {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 938.045331] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 938.049306] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-baef9a30-f1ea-4ab4-8a5c-a9205c64af77 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.070732] env[61906]: DEBUG oslo_vmware.rw_handles [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 938.070732] env[61906]: value = "vm-284797" [ 938.070732] env[61906]: _type = "VirtualMachine" [ 938.070732] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 938.070732] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-bd1f7286-00b0-4b73-88b4-f1a20f4d9b15 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.077950] env[61906]: DEBUG oslo_vmware.rw_handles [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lease: (returnval){ [ 938.077950] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]520c3ae5-5365-ac73-4b0c-d69c7dbd9649" [ 938.077950] env[61906]: _type = "HttpNfcLease" [ 938.077950] env[61906]: } obtained for exporting VM: (result){ [ 938.077950] env[61906]: value = "vm-284797" [ 938.077950] env[61906]: _type = "VirtualMachine" [ 938.077950] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 938.077950] env[61906]: DEBUG oslo_vmware.api [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the lease: (returnval){ [ 938.077950] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]520c3ae5-5365-ac73-4b0c-d69c7dbd9649" [ 938.077950] env[61906]: _type = "HttpNfcLease" [ 938.077950] env[61906]: } to be ready. {{(pid=61906) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 938.084149] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 938.084437] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 938.084668] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Deleting the datastore file [datastore2] 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 938.086383] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7d376c92-f539-4d48-a2c2-ac6edcb9a811 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.088427] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 938.088427] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]520c3ae5-5365-ac73-4b0c-d69c7dbd9649" [ 938.088427] env[61906]: _type = "HttpNfcLease" [ 938.088427] env[61906]: } is initializing. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 938.095523] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 938.095523] env[61906]: value = "task-1333401" [ 938.095523] env[61906]: _type = "Task" [ 938.095523] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.107664] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333401, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.486480] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.951s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.487136] env[61906]: DEBUG nova.compute.manager [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 938.489997] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 4.695s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.499531] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "refresh_cache-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.505378] env[61906]: DEBUG nova.compute.manager [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 938.532212] env[61906]: DEBUG nova.virt.hardware [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 938.532565] env[61906]: DEBUG nova.virt.hardware [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 938.532751] env[61906]: DEBUG nova.virt.hardware [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 938.533013] env[61906]: DEBUG nova.virt.hardware [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 938.533217] env[61906]: DEBUG nova.virt.hardware [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 938.533350] env[61906]: DEBUG nova.virt.hardware [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 938.533560] env[61906]: DEBUG nova.virt.hardware [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 938.533724] env[61906]: DEBUG nova.virt.hardware [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 938.533890] env[61906]: DEBUG nova.virt.hardware [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 938.534080] env[61906]: DEBUG nova.virt.hardware [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 938.534246] env[61906]: DEBUG nova.virt.hardware [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 938.539549] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Reconfiguring VM instance instance-0000004b to detach disk 2000 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 938.540927] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e6d0d2c-44c0-49cd-a135-4496f1ac8e38 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.556560] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.563847] env[61906]: DEBUG oslo_vmware.api [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 938.563847] env[61906]: value = "task-1333403" [ 938.563847] env[61906]: _type = "Task" [ 938.563847] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.577417] env[61906]: DEBUG oslo_vmware.api [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333403, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.584948] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 938.584948] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]520c3ae5-5365-ac73-4b0c-d69c7dbd9649" [ 938.584948] env[61906]: _type = "HttpNfcLease" [ 938.584948] env[61906]: } is ready. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 938.586063] env[61906]: DEBUG oslo_vmware.rw_handles [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 938.586063] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]520c3ae5-5365-ac73-4b0c-d69c7dbd9649" [ 938.586063] env[61906]: _type = "HttpNfcLease" [ 938.586063] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 938.587832] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccf1d8a5-a41c-435a-9440-543f23a76e2e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.596642] env[61906]: DEBUG oslo_vmware.rw_handles [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ddb914-cd54-f253-9ba3-0e7b07f9b0c9/disk-0.vmdk from lease info. {{(pid=61906) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 938.597133] env[61906]: DEBUG oslo_vmware.rw_handles [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ddb914-cd54-f253-9ba3-0e7b07f9b0c9/disk-0.vmdk for reading. {{(pid=61906) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 938.663979] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333401, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.122975} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.665618] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 938.665845] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 938.666009] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 938.698924] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1678ed99-cdff-441f-a89b-8fcb66542f67 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.858640] env[61906]: DEBUG nova.network.neutron [req-47754955-eb75-4e5c-8bd7-a7669509fd8b req-ab1f0894-3d33-4d09-a65d-347c4c0b3699 service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Added VIF to instance network info cache for port 4f8ab1c1-fd91-48d3-90e0-ed5ff9fa0b4d. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3546}} [ 938.859147] env[61906]: DEBUG nova.network.neutron [req-47754955-eb75-4e5c-8bd7-a7669509fd8b req-ab1f0894-3d33-4d09-a65d-347c4c0b3699 service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Updating instance_info_cache with network_info: [{"id": "3e04c5c7-8d7a-46ae-a2d7-52dc429650f9", "address": "fa:16:3e:f1:15:76", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e04c5c7-8d", "ovs_interfaceid": "3e04c5c7-8d7a-46ae-a2d7-52dc429650f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c2232d91-1c2c-4a66-9026-933169da90e1", "address": "fa:16:3e:80:0d:49", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2232d91-1c", "ovs_interfaceid": "c2232d91-1c2c-4a66-9026-933169da90e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "4f8ab1c1-fd91-48d3-90e0-ed5ff9fa0b4d", "address": "fa:16:3e:ee:f2:da", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f8ab1c1-fd", "ovs_interfaceid": "4f8ab1c1-fd91-48d3-90e0-ed5ff9fa0b4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.993421] env[61906]: DEBUG nova.compute.utils [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 938.995194] env[61906]: DEBUG nova.compute.manager [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 938.995539] env[61906]: DEBUG nova.network.neutron [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 939.030943] env[61906]: DEBUG oslo_concurrency.lockutils [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.047072] env[61906]: DEBUG nova.policy [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e89d45df95134c709b569d06744dea5e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1b33dea0794a48f78b9f519cb269a8c2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 939.075177] env[61906]: DEBUG oslo_vmware.api [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333403, 'name': ReconfigVM_Task, 'duration_secs': 0.229442} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.075996] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Reconfigured VM instance instance-0000004b to detach disk 2000 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 939.076825] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6505e6e6-3238-49a1-8c68-a3ce747495a1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.100720] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15/e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 939.101255] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a3a1ab9d-960a-440e-8e22-002523ed24dd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.121176] env[61906]: DEBUG oslo_vmware.api [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 939.121176] env[61906]: value = "task-1333404" [ 939.121176] env[61906]: _type = "Task" [ 939.121176] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.134344] env[61906]: DEBUG oslo_vmware.api [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333404, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.331407] env[61906]: DEBUG nova.network.neutron [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Successfully created port: fb758a9b-f65f-4d8c-84a3-290128c68110 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 939.362887] env[61906]: DEBUG oslo_concurrency.lockutils [req-47754955-eb75-4e5c-8bd7-a7669509fd8b req-ab1f0894-3d33-4d09-a65d-347c4c0b3699 service nova] Releasing lock "refresh_cache-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.366067] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired lock "refresh_cache-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.366400] env[61906]: DEBUG nova.network.neutron [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 939.506608] env[61906]: DEBUG nova.compute.manager [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 939.512309] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Applying migration context for instance e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15 as it has an incoming, in-progress migration f037f355-83fb-4668-9837-81e957b39c1e. Migration status is migrating {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 939.515061] env[61906]: INFO nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Updating resource usage from migration f037f355-83fb-4668-9837-81e957b39c1e [ 939.538180] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 939.539720] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 7504dfa6-bedc-4701-b4fc-60e19e742276 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 939.539720] env[61906]: WARNING nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance da493512-d996-4de7-9f47-cadcbc4fbcb9 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 939.539720] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance b1159533-c970-49d6-af42-b954b20d92fb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 939.539720] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance ad2db007-e7aa-4c66-9020-ac4acfcff25a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 939.540171] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 1a335019-915b-4065-b22f-d2f0af728d6f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 939.540306] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 11cb9d89-2304-45a6-b4bb-fcf8417a0518 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 939.540581] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 939.541165] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Migration f037f355-83fb-4668-9837-81e957b39c1e is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 939.541165] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 939.541268] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 83b59194-b86f-416a-a7fd-18f75551c2c0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 939.632588] env[61906]: DEBUG oslo_vmware.api [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333404, 'name': ReconfigVM_Task, 'duration_secs': 0.460782} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.633263] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Reconfigured VM instance instance-0000004b to attach disk [datastore1] e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15/e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 939.633607] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Updating instance 'e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15' progress to 50 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 939.704009] env[61906]: DEBUG nova.virt.hardware [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 939.704765] env[61906]: DEBUG nova.virt.hardware [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 939.704765] env[61906]: DEBUG nova.virt.hardware [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 939.704765] env[61906]: DEBUG nova.virt.hardware [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 939.704921] env[61906]: DEBUG nova.virt.hardware [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 939.705072] env[61906]: DEBUG nova.virt.hardware [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 939.705286] env[61906]: DEBUG nova.virt.hardware [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 939.705494] env[61906]: DEBUG nova.virt.hardware [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 939.705626] env[61906]: DEBUG nova.virt.hardware [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 939.705783] env[61906]: DEBUG nova.virt.hardware [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 939.706047] env[61906]: DEBUG nova.virt.hardware [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 939.706953] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23bc5f8c-8e5f-4aec-996b-41db1e88ffe1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.715761] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcc0e3bd-be00-4a10-aab0-dbbac9ef7021 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.730080] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Instance VIF info [] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 939.735825] env[61906]: DEBUG oslo.service.loopingcall [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 939.736355] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 939.736550] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3d48bb05-0fe8-4a69-80b2-ef799f4f9701 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.755974] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 939.755974] env[61906]: value = "task-1333405" [ 939.755974] env[61906]: _type = "Task" [ 939.755974] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.769711] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333405, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.910311] env[61906]: WARNING nova.network.neutron [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] 37d8737f-db44-4cf6-8e34-31b4cbd2f82e already exists in list: networks containing: ['37d8737f-db44-4cf6-8e34-31b4cbd2f82e']. ignoring it [ 939.910630] env[61906]: WARNING nova.network.neutron [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] 37d8737f-db44-4cf6-8e34-31b4cbd2f82e already exists in list: networks containing: ['37d8737f-db44-4cf6-8e34-31b4cbd2f82e']. ignoring it [ 939.910889] env[61906]: WARNING nova.network.neutron [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] 37d8737f-db44-4cf6-8e34-31b4cbd2f82e already exists in list: networks containing: ['37d8737f-db44-4cf6-8e34-31b4cbd2f82e']. ignoring it [ 939.911168] env[61906]: WARNING nova.network.neutron [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] 4f8ab1c1-fd91-48d3-90e0-ed5ff9fa0b4d already exists in list: port_ids containing: ['4f8ab1c1-fd91-48d3-90e0-ed5ff9fa0b4d']. ignoring it [ 940.044886] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 0f846f95-2573-4450-9918-a34467d73363 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 940.045321] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 940.045321] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2496MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 940.144113] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f6c6e81-3f10-4a12-8c19-10d4a65261c6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.173679] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3fc96cd-e183-4059-95b0-93f838f0b4b6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.193419] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Updating instance 'e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15' progress to 67 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 940.245073] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9be66530-b35e-4807-85a3-b545a3c07ae0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.254149] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-608fd377-68bc-4240-ba11-b15f244e8ab4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.268861] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333405, 'name': CreateVM_Task, 'duration_secs': 0.402413} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.295225] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 940.296296] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.296484] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.297119] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 940.297898] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12e1afe0-8353-4eb3-b5e8-73ae6b0eb95f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.300783] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68290246-ca82-429b-bda5-bab29dab1b86 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.305994] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 940.305994] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]529bda8a-4d5a-a32d-97c7-d12e3ee4e67c" [ 940.305994] env[61906]: _type = "Task" [ 940.305994] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.312448] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-667acdcf-fd76-4541-a00c-ee7ae566be30 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.326498] env[61906]: DEBUG nova.compute.provider_tree [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 940.333067] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]529bda8a-4d5a-a32d-97c7-d12e3ee4e67c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.409265] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-999a1791-dfd9-4a13-9a0d-d223f011b2c5 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Volume attach. Driver type: vmdk {{(pid=61906) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 940.409639] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-999a1791-dfd9-4a13-9a0d-d223f011b2c5 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284799', 'volume_id': 'a8641678-cc5c-458c-a0a6-daf5317bf7d7', 'name': 'volume-a8641678-cc5c-458c-a0a6-daf5317bf7d7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7504dfa6-bedc-4701-b4fc-60e19e742276', 'attached_at': '', 'detached_at': '', 'volume_id': 'a8641678-cc5c-458c-a0a6-daf5317bf7d7', 'serial': 'a8641678-cc5c-458c-a0a6-daf5317bf7d7'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 940.410547] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0b43081-4fa7-4d90-9b16-73084f7e2c10 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.428638] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dea32c9-78cd-4d08-a9a3-b3303941339c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.431571] env[61906]: DEBUG nova.network.neutron [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Updating instance_info_cache with network_info: [{"id": "3e04c5c7-8d7a-46ae-a2d7-52dc429650f9", "address": "fa:16:3e:f1:15:76", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e04c5c7-8d", "ovs_interfaceid": "3e04c5c7-8d7a-46ae-a2d7-52dc429650f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c2232d91-1c2c-4a66-9026-933169da90e1", "address": "fa:16:3e:80:0d:49", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2232d91-1c", "ovs_interfaceid": "c2232d91-1c2c-4a66-9026-933169da90e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "4f8ab1c1-fd91-48d3-90e0-ed5ff9fa0b4d", "address": "fa:16:3e:ee:f2:da", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f8ab1c1-fd", "ovs_interfaceid": "4f8ab1c1-fd91-48d3-90e0-ed5ff9fa0b4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.456938] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-999a1791-dfd9-4a13-9a0d-d223f011b2c5 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] volume-a8641678-cc5c-458c-a0a6-daf5317bf7d7/volume-a8641678-cc5c-458c-a0a6-daf5317bf7d7.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 940.457990] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d5e9a50b-af28-496f-890b-584deb1bbacb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.477645] env[61906]: DEBUG oslo_vmware.api [None req-999a1791-dfd9-4a13-9a0d-d223f011b2c5 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 940.477645] env[61906]: value = "task-1333406" [ 940.477645] env[61906]: _type = "Task" [ 940.477645] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.488826] env[61906]: DEBUG oslo_vmware.api [None req-999a1791-dfd9-4a13-9a0d-d223f011b2c5 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333406, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.518067] env[61906]: DEBUG nova.compute.manager [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 940.544120] env[61906]: DEBUG nova.virt.hardware [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 940.544413] env[61906]: DEBUG nova.virt.hardware [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 940.544582] env[61906]: DEBUG nova.virt.hardware [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 940.544787] env[61906]: DEBUG nova.virt.hardware [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 940.544934] env[61906]: DEBUG nova.virt.hardware [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 940.545128] env[61906]: DEBUG nova.virt.hardware [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 940.545351] env[61906]: DEBUG nova.virt.hardware [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 940.545575] env[61906]: DEBUG nova.virt.hardware [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 940.545774] env[61906]: DEBUG nova.virt.hardware [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 940.545951] env[61906]: DEBUG nova.virt.hardware [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 940.546215] env[61906]: DEBUG nova.virt.hardware [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 940.547139] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4211754-f0f7-4eaa-a957-caf37c505a26 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.556633] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4422b5f-e38b-441f-9826-ab5f87af9d5f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.737785] env[61906]: DEBUG nova.network.neutron [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Port b54c27f8-7727-4166-af82-5b6fe0d3519e binding to destination host cpu-1 is already ACTIVE {{(pid=61906) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 940.820809] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]529bda8a-4d5a-a32d-97c7-d12e3ee4e67c, 'name': SearchDatastore_Task, 'duration_secs': 0.024518} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.820809] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.820809] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 940.821085] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.821179] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.821359] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 940.821642] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-602317ac-8fb6-498a-afb3-071ed96dad92 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.835121] env[61906]: DEBUG nova.scheduler.client.report [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 940.839291] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 940.839607] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 940.840513] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2233a6a8-ed07-4daf-8245-960b855636e6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.847529] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 940.847529] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]526126ee-372d-ab29-de85-3603e3a8d0b0" [ 940.847529] env[61906]: _type = "Task" [ 940.847529] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.855819] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]526126ee-372d-ab29-de85-3603e3a8d0b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.870439] env[61906]: DEBUG nova.compute.manager [req-19865401-7e1e-468f-a9a4-2ee9656ffa2f req-4245fdaf-5d4e-4709-84d3-5bf1eb8b0c8a service nova] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Received event network-vif-plugged-fb758a9b-f65f-4d8c-84a3-290128c68110 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 940.870661] env[61906]: DEBUG oslo_concurrency.lockutils [req-19865401-7e1e-468f-a9a4-2ee9656ffa2f req-4245fdaf-5d4e-4709-84d3-5bf1eb8b0c8a service nova] Acquiring lock "83b59194-b86f-416a-a7fd-18f75551c2c0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.870902] env[61906]: DEBUG oslo_concurrency.lockutils [req-19865401-7e1e-468f-a9a4-2ee9656ffa2f req-4245fdaf-5d4e-4709-84d3-5bf1eb8b0c8a service nova] Lock "83b59194-b86f-416a-a7fd-18f75551c2c0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.871043] env[61906]: DEBUG oslo_concurrency.lockutils [req-19865401-7e1e-468f-a9a4-2ee9656ffa2f req-4245fdaf-5d4e-4709-84d3-5bf1eb8b0c8a service nova] Lock "83b59194-b86f-416a-a7fd-18f75551c2c0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.871221] env[61906]: DEBUG nova.compute.manager [req-19865401-7e1e-468f-a9a4-2ee9656ffa2f req-4245fdaf-5d4e-4709-84d3-5bf1eb8b0c8a service nova] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] No waiting events found dispatching network-vif-plugged-fb758a9b-f65f-4d8c-84a3-290128c68110 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 940.871389] env[61906]: WARNING nova.compute.manager [req-19865401-7e1e-468f-a9a4-2ee9656ffa2f req-4245fdaf-5d4e-4709-84d3-5bf1eb8b0c8a service nova] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Received unexpected event network-vif-plugged-fb758a9b-f65f-4d8c-84a3-290128c68110 for instance with vm_state building and task_state spawning. [ 940.934899] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Releasing lock "refresh_cache-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.935628] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.935803] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.936985] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf2fa887-fcab-4293-aa9a-aee41fdff671 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.956144] env[61906]: DEBUG nova.virt.hardware [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 940.956400] env[61906]: DEBUG nova.virt.hardware [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 940.956565] env[61906]: DEBUG nova.virt.hardware [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 940.956752] env[61906]: DEBUG nova.virt.hardware [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 940.956904] env[61906]: DEBUG nova.virt.hardware [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 940.957076] env[61906]: DEBUG nova.virt.hardware [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 940.957291] env[61906]: DEBUG nova.virt.hardware [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 940.957455] env[61906]: DEBUG nova.virt.hardware [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 940.957627] env[61906]: DEBUG nova.virt.hardware [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 940.957794] env[61906]: DEBUG nova.virt.hardware [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 940.957970] env[61906]: DEBUG nova.virt.hardware [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 940.964780] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Reconfiguring VM to attach interface {{(pid=61906) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 940.965142] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-97c47ff4-1c48-4777-b4ad-e7b6a064f20a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.977547] env[61906]: DEBUG nova.network.neutron [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Successfully updated port: fb758a9b-f65f-4d8c-84a3-290128c68110 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 940.985649] env[61906]: DEBUG oslo_vmware.api [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 940.985649] env[61906]: value = "task-1333407" [ 940.985649] env[61906]: _type = "Task" [ 940.985649] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.989818] env[61906]: DEBUG oslo_vmware.api [None req-999a1791-dfd9-4a13-9a0d-d223f011b2c5 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333406, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.997511] env[61906]: DEBUG oslo_vmware.api [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333407, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.342210] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 941.342628] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.853s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.342908] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.786s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.343178] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.345485] env[61906]: DEBUG oslo_concurrency.lockutils [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.315s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.347056] env[61906]: INFO nova.compute.claims [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 941.360347] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]526126ee-372d-ab29-de85-3603e3a8d0b0, 'name': SearchDatastore_Task, 'duration_secs': 0.015895} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.361281] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-458e5343-ebf8-41ad-9999-cd7d4506dc8a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.366857] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 941.366857] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52834bc3-0aa2-1dd5-65b6-8307b4828743" [ 941.366857] env[61906]: _type = "Task" [ 941.366857] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.371097] env[61906]: INFO nova.scheduler.client.report [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Deleted allocations for instance da493512-d996-4de7-9f47-cadcbc4fbcb9 [ 941.375986] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52834bc3-0aa2-1dd5-65b6-8307b4828743, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.482654] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "refresh_cache-83b59194-b86f-416a-a7fd-18f75551c2c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.482751] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquired lock "refresh_cache-83b59194-b86f-416a-a7fd-18f75551c2c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.483032] env[61906]: DEBUG nova.network.neutron [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 941.491200] env[61906]: DEBUG oslo_vmware.api [None req-999a1791-dfd9-4a13-9a0d-d223f011b2c5 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333406, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.500772] env[61906]: DEBUG oslo_vmware.api [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333407, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.759864] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.760076] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.760268] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.878454] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52834bc3-0aa2-1dd5-65b6-8307b4828743, 'name': SearchDatastore_Task, 'duration_secs': 0.010929} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.878760] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.879043] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5/8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 941.881245] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-40bc3bf3-0b7c-4aa1-b800-0dd932a55c49 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.883608] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa0508e8-7d66-4d56-87e2-41589c1fd58e tempest-ServersTestBootFromVolume-1160277877 tempest-ServersTestBootFromVolume-1160277877-project-member] Lock "da493512-d996-4de7-9f47-cadcbc4fbcb9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.960s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.888629] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 941.888629] env[61906]: value = "task-1333408" [ 941.888629] env[61906]: _type = "Task" [ 941.888629] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.896942] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333408, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.989750] env[61906]: DEBUG oslo_vmware.api [None req-999a1791-dfd9-4a13-9a0d-d223f011b2c5 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333406, 'name': ReconfigVM_Task, 'duration_secs': 1.357848} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.990078] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-999a1791-dfd9-4a13-9a0d-d223f011b2c5 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Reconfigured VM instance instance-00000046 to attach disk [datastore2] volume-a8641678-cc5c-458c-a0a6-daf5317bf7d7/volume-a8641678-cc5c-458c-a0a6-daf5317bf7d7.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 941.995084] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fbd1d818-f412-43c2-aeda-6fde9e014d6f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.014779] env[61906]: DEBUG oslo_vmware.api [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333407, 'name': ReconfigVM_Task, 'duration_secs': 0.587088} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.016365] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Releasing lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.016599] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Reconfigured VM to attach interface {{(pid=61906) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 942.019168] env[61906]: DEBUG oslo_vmware.api [None req-999a1791-dfd9-4a13-9a0d-d223f011b2c5 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 942.019168] env[61906]: value = "task-1333409" [ 942.019168] env[61906]: _type = "Task" [ 942.019168] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.024085] env[61906]: DEBUG nova.network.neutron [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 942.035840] env[61906]: DEBUG oslo_vmware.api [None req-999a1791-dfd9-4a13-9a0d-d223f011b2c5 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333409, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.206489] env[61906]: DEBUG nova.network.neutron [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Updating instance_info_cache with network_info: [{"id": "fb758a9b-f65f-4d8c-84a3-290128c68110", "address": "fa:16:3e:29:18:30", "network": {"id": "ddc08dc9-d08f-4fb9-9bee-1a030ed69f76", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1124067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b33dea0794a48f78b9f519cb269a8c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f1b07b1-e4e5-4842-9090-07fb2c3e124b", "external-id": "nsx-vlan-transportzone-646", "segmentation_id": 646, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb758a9b-f6", "ovs_interfaceid": "fb758a9b-f65f-4d8c-84a3-290128c68110", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.403160] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333408, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.521556] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8a67be17-f4da-4b62-ac32-39d26b228d81 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "interface-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8-4f8ab1c1-fd91-48d3-90e0-ed5ff9fa0b4d" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.583s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.525091] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57dd5724-11f7-4455-bf0a-69a64c35f2d4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.538272] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c5d1347-fa0b-46bb-95a9-cd8d355db42b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.542021] env[61906]: DEBUG oslo_vmware.api [None req-999a1791-dfd9-4a13-9a0d-d223f011b2c5 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333409, 'name': ReconfigVM_Task, 'duration_secs': 0.156135} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.542423] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-999a1791-dfd9-4a13-9a0d-d223f011b2c5 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284799', 'volume_id': 'a8641678-cc5c-458c-a0a6-daf5317bf7d7', 'name': 'volume-a8641678-cc5c-458c-a0a6-daf5317bf7d7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7504dfa6-bedc-4701-b4fc-60e19e742276', 'attached_at': '', 'detached_at': '', 'volume_id': 'a8641678-cc5c-458c-a0a6-daf5317bf7d7', 'serial': 'a8641678-cc5c-458c-a0a6-daf5317bf7d7'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 942.572151] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5f8ed8a-d816-464b-84f8-4b68828e4187 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.580521] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4264f7df-01c0-4889-b70c-d9386ce8249f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.596466] env[61906]: DEBUG nova.compute.provider_tree [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 942.709249] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Releasing lock "refresh_cache-83b59194-b86f-416a-a7fd-18f75551c2c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.709599] env[61906]: DEBUG nova.compute.manager [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Instance network_info: |[{"id": "fb758a9b-f65f-4d8c-84a3-290128c68110", "address": "fa:16:3e:29:18:30", "network": {"id": "ddc08dc9-d08f-4fb9-9bee-1a030ed69f76", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1124067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b33dea0794a48f78b9f519cb269a8c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f1b07b1-e4e5-4842-9090-07fb2c3e124b", "external-id": "nsx-vlan-transportzone-646", "segmentation_id": 646, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb758a9b-f6", "ovs_interfaceid": "fb758a9b-f65f-4d8c-84a3-290128c68110", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 942.710177] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:29:18:30', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6f1b07b1-e4e5-4842-9090-07fb2c3e124b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fb758a9b-f65f-4d8c-84a3-290128c68110', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 942.717656] env[61906]: DEBUG oslo.service.loopingcall [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 942.717909] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 942.718157] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-abc8fe9e-fdeb-44ef-9d6c-8be832fb75b9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.739171] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 942.739171] env[61906]: value = "task-1333410" [ 942.739171] env[61906]: _type = "Task" [ 942.739171] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.747455] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333410, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.838617] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "refresh_cache-e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.838849] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired lock "refresh_cache-e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.839049] env[61906]: DEBUG nova.network.neutron [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 942.900529] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333408, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.53826} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.900832] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5/8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 942.901118] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 942.901398] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-781d3a18-7cba-4e3f-bdfc-222b0cae046f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.908457] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 942.908457] env[61906]: value = "task-1333411" [ 942.908457] env[61906]: _type = "Task" [ 942.908457] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.916444] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333411, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.099461] env[61906]: DEBUG nova.scheduler.client.report [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 943.251784] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333410, 'name': CreateVM_Task, 'duration_secs': 0.428561} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.251963] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 943.252714] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.252964] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.253310] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 943.253871] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ff204d7-f301-4b59-af78-364e6d3b0fff {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.258551] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 943.258551] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52759c92-a11c-bd0f-e52f-b0670c1df42d" [ 943.258551] env[61906]: _type = "Task" [ 943.258551] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.267281] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52759c92-a11c-bd0f-e52f-b0670c1df42d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.293514] env[61906]: DEBUG nova.compute.manager [req-18e14d07-2db8-4fd2-99e6-300a8c4019dd req-10b19cc3-2ce1-4580-a233-8cbaabfb298a service nova] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Received event network-changed-fb758a9b-f65f-4d8c-84a3-290128c68110 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.293759] env[61906]: DEBUG nova.compute.manager [req-18e14d07-2db8-4fd2-99e6-300a8c4019dd req-10b19cc3-2ce1-4580-a233-8cbaabfb298a service nova] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Refreshing instance network info cache due to event network-changed-fb758a9b-f65f-4d8c-84a3-290128c68110. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 943.293975] env[61906]: DEBUG oslo_concurrency.lockutils [req-18e14d07-2db8-4fd2-99e6-300a8c4019dd req-10b19cc3-2ce1-4580-a233-8cbaabfb298a service nova] Acquiring lock "refresh_cache-83b59194-b86f-416a-a7fd-18f75551c2c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.296013] env[61906]: DEBUG oslo_concurrency.lockutils [req-18e14d07-2db8-4fd2-99e6-300a8c4019dd req-10b19cc3-2ce1-4580-a233-8cbaabfb298a service nova] Acquired lock "refresh_cache-83b59194-b86f-416a-a7fd-18f75551c2c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.296252] env[61906]: DEBUG nova.network.neutron [req-18e14d07-2db8-4fd2-99e6-300a8c4019dd req-10b19cc3-2ce1-4580-a233-8cbaabfb298a service nova] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Refreshing network info cache for port fb758a9b-f65f-4d8c-84a3-290128c68110 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 943.421416] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333411, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079273} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.421740] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 943.424969] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c58fa55-2106-4e78-8db5-4dff0644e0a8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.450076] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5/8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 943.450076] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c69b6dcd-6e3b-4fa1-b0bc-a97583dc2d59 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.472696] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 943.472696] env[61906]: value = "task-1333412" [ 943.472696] env[61906]: _type = "Task" [ 943.472696] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.480068] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333412, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.586397] env[61906]: DEBUG nova.objects.instance [None req-999a1791-dfd9-4a13-9a0d-d223f011b2c5 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lazy-loading 'flavor' on Instance uuid 7504dfa6-bedc-4701-b4fc-60e19e742276 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 943.604655] env[61906]: DEBUG oslo_concurrency.lockutils [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.259s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.605327] env[61906]: DEBUG nova.compute.manager [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 943.769221] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52759c92-a11c-bd0f-e52f-b0670c1df42d, 'name': SearchDatastore_Task, 'duration_secs': 0.009422} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.769554] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.769800] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 943.770057] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.770212] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.770400] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 943.770680] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-67d40696-54b1-4fbb-9484-8a8a778e9213 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.779312] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 943.779504] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 943.780272] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b8ec854-306b-478a-bea8-58af7240bedb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.786458] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 943.786458] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52f124e1-20a5-295a-abcb-0cd15aec7d81" [ 943.786458] env[61906]: _type = "Task" [ 943.786458] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.791550] env[61906]: DEBUG nova.network.neutron [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Updating instance_info_cache with network_info: [{"id": "b54c27f8-7727-4166-af82-5b6fe0d3519e", "address": "fa:16:3e:ae:b6:e5", "network": {"id": "af039bdd-ffd2-483d-9655-e7df48936290", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-111531540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f7fc2042e6549cabe98838990f35842", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb54c27f8-77", "ovs_interfaceid": "b54c27f8-7727-4166-af82-5b6fe0d3519e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.804738] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52f124e1-20a5-295a-abcb-0cd15aec7d81, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.986673] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333412, 'name': ReconfigVM_Task, 'duration_secs': 0.270508} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.986673] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5/8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 943.986673] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bbb128e0-6764-4312-831d-90daaf84778c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.993186] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 943.993186] env[61906]: value = "task-1333413" [ 943.993186] env[61906]: _type = "Task" [ 943.993186] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.004098] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333413, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.096836] env[61906]: DEBUG oslo_concurrency.lockutils [None req-999a1791-dfd9-4a13-9a0d-d223f011b2c5 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "7504dfa6-bedc-4701-b4fc-60e19e742276" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.315s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.112906] env[61906]: DEBUG nova.compute.utils [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 944.115321] env[61906]: DEBUG nova.compute.manager [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 944.115488] env[61906]: DEBUG nova.network.neutron [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 944.187706] env[61906]: DEBUG nova.network.neutron [req-18e14d07-2db8-4fd2-99e6-300a8c4019dd req-10b19cc3-2ce1-4580-a233-8cbaabfb298a service nova] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Updated VIF entry in instance network info cache for port fb758a9b-f65f-4d8c-84a3-290128c68110. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 944.188103] env[61906]: DEBUG nova.network.neutron [req-18e14d07-2db8-4fd2-99e6-300a8c4019dd req-10b19cc3-2ce1-4580-a233-8cbaabfb298a service nova] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Updating instance_info_cache with network_info: [{"id": "fb758a9b-f65f-4d8c-84a3-290128c68110", "address": "fa:16:3e:29:18:30", "network": {"id": "ddc08dc9-d08f-4fb9-9bee-1a030ed69f76", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1124067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b33dea0794a48f78b9f519cb269a8c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f1b07b1-e4e5-4842-9090-07fb2c3e124b", "external-id": "nsx-vlan-transportzone-646", "segmentation_id": 646, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb758a9b-f6", "ovs_interfaceid": "fb758a9b-f65f-4d8c-84a3-290128c68110", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.217353] env[61906]: DEBUG nova.policy [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b49e5a6aeb1d43e290a645e14861b889', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '53644a4143a24759a9ff2b5e28b84fb4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 944.298425] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Releasing lock "refresh_cache-e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.302228] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52f124e1-20a5-295a-abcb-0cd15aec7d81, 'name': SearchDatastore_Task, 'duration_secs': 0.010532} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.303096] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d6b68df-b114-4e67-a909-e281104feb99 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.310162] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 944.310162] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52f665f9-6baf-cd88-b3b7-9ce339657c67" [ 944.310162] env[61906]: _type = "Task" [ 944.310162] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.317743] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52f665f9-6baf-cd88-b3b7-9ce339657c67, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.445204] env[61906]: DEBUG oslo_concurrency.lockutils [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "interface-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8-c2232d91-1c2c-4a66-9026-933169da90e1" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.445635] env[61906]: DEBUG oslo_concurrency.lockutils [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "interface-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8-c2232d91-1c2c-4a66-9026-933169da90e1" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.502745] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333413, 'name': Rename_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.619820] env[61906]: DEBUG nova.compute.manager [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 944.691351] env[61906]: DEBUG oslo_concurrency.lockutils [req-18e14d07-2db8-4fd2-99e6-300a8c4019dd req-10b19cc3-2ce1-4580-a233-8cbaabfb298a service nova] Releasing lock "refresh_cache-83b59194-b86f-416a-a7fd-18f75551c2c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.766449] env[61906]: DEBUG nova.network.neutron [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Successfully created port: 12569035-8b00-43d3-bc6d-5a22464fb53d {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 944.828509] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52f665f9-6baf-cd88-b3b7-9ce339657c67, 'name': SearchDatastore_Task, 'duration_secs': 0.024503} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.829522] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.829522] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 83b59194-b86f-416a-a7fd-18f75551c2c0/83b59194-b86f-416a-a7fd-18f75551c2c0.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 944.829522] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eddb57aa-161b-473a-8c01-f31d24aae2b8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.836789] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 944.836789] env[61906]: value = "task-1333414" [ 944.836789] env[61906]: _type = "Task" [ 944.836789] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.843274] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9947f169-57ed-4ccd-a7de-c123d0c52b12 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.848669] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333414, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.867339] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa46470d-8b3c-49d9-9e73-41f8c746000d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.872062] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Updating instance 'e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15' progress to 83 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 944.949226] env[61906]: DEBUG oslo_concurrency.lockutils [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.949458] env[61906]: DEBUG oslo_concurrency.lockutils [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.950772] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df1d9fc4-3803-486b-9156-ac2ff69454a0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.970865] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b245eed-cc5d-4088-bd6b-7cb75d482cc0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.000953] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Reconfiguring VM to detach interface {{(pid=61906) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 945.004653] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae9b159a-66e5-4019-9dca-967485a29c14 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.034200] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333413, 'name': Rename_Task, 'duration_secs': 0.907338} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.035878] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 945.036255] env[61906]: DEBUG oslo_vmware.api [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 945.036255] env[61906]: value = "task-1333415" [ 945.036255] env[61906]: _type = "Task" [ 945.036255] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.036462] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-18c8eb62-1c8f-470d-8736-35b635956c2c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.050467] env[61906]: DEBUG oslo_vmware.api [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333415, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.052444] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 945.052444] env[61906]: value = "task-1333416" [ 945.052444] env[61906]: _type = "Task" [ 945.052444] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.061204] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333416, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.347643] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333414, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.378395] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 945.378758] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bf3f60fb-fc95-403d-ba5f-1f0ea860c8b5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.389030] env[61906]: DEBUG oslo_vmware.api [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 945.389030] env[61906]: value = "task-1333417" [ 945.389030] env[61906]: _type = "Task" [ 945.389030] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.395676] env[61906]: DEBUG oslo_vmware.api [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333417, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.527806] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0855be2e-4c0a-4562-a4e3-26755435a6fe tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "b1159533-c970-49d6-af42-b954b20d92fb" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.528230] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0855be2e-4c0a-4562-a4e3-26755435a6fe tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "b1159533-c970-49d6-af42-b954b20d92fb" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.549545] env[61906]: DEBUG oslo_vmware.api [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333415, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.562334] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333416, 'name': PowerOnVM_Task} progress is 90%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.628932] env[61906]: DEBUG nova.compute.manager [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 945.656055] env[61906]: DEBUG nova.virt.hardware [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 945.656333] env[61906]: DEBUG nova.virt.hardware [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 945.656498] env[61906]: DEBUG nova.virt.hardware [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 945.656691] env[61906]: DEBUG nova.virt.hardware [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 945.656842] env[61906]: DEBUG nova.virt.hardware [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 945.656994] env[61906]: DEBUG nova.virt.hardware [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 945.657267] env[61906]: DEBUG nova.virt.hardware [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 945.657443] env[61906]: DEBUG nova.virt.hardware [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 945.657621] env[61906]: DEBUG nova.virt.hardware [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 945.657820] env[61906]: DEBUG nova.virt.hardware [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 945.658075] env[61906]: DEBUG nova.virt.hardware [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 945.659106] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6fa4e96-d676-4c7f-ae33-f5453752be4c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.669218] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d2a4afc-8fa4-40b7-bc19-853e9a6e1749 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.849834] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333414, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.550348} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.850234] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 83b59194-b86f-416a-a7fd-18f75551c2c0/83b59194-b86f-416a-a7fd-18f75551c2c0.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 945.850461] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 945.850753] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-59a07c12-2396-4263-bfc4-c92f75e3f908 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.857762] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 945.857762] env[61906]: value = "task-1333418" [ 945.857762] env[61906]: _type = "Task" [ 945.857762] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.866424] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333418, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.897348] env[61906]: DEBUG oslo_vmware.api [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333417, 'name': PowerOnVM_Task, 'duration_secs': 0.436602} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.897720] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 945.898014] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d0af8c7c-0ce7-4b55-997f-3a1b84bb21e0 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Updating instance 'e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15' progress to 100 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 946.034688] env[61906]: DEBUG nova.compute.utils [None req-0855be2e-4c0a-4562-a4e3-26755435a6fe tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 946.051684] env[61906]: DEBUG oslo_vmware.api [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333415, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.062296] env[61906]: DEBUG oslo_vmware.api [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333416, 'name': PowerOnVM_Task, 'duration_secs': 0.571357} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.062611] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 946.062823] env[61906]: DEBUG nova.compute.manager [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 946.063632] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5754d09d-742e-4691-ad95-82912e99706f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.271261] env[61906]: DEBUG oslo_vmware.rw_handles [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ddb914-cd54-f253-9ba3-0e7b07f9b0c9/disk-0.vmdk. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 946.272252] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbef406a-383c-4cfb-8b67-2b427ed0a2de {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.278727] env[61906]: DEBUG oslo_vmware.rw_handles [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ddb914-cd54-f253-9ba3-0e7b07f9b0c9/disk-0.vmdk is in state: ready. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 946.278903] env[61906]: ERROR oslo_vmware.rw_handles [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ddb914-cd54-f253-9ba3-0e7b07f9b0c9/disk-0.vmdk due to incomplete transfer. [ 946.279147] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b3016b7e-8d1a-4bf4-9a74-750206bab49b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.286544] env[61906]: DEBUG oslo_vmware.rw_handles [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ddb914-cd54-f253-9ba3-0e7b07f9b0c9/disk-0.vmdk. {{(pid=61906) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 946.286761] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Uploaded image 670311c4-3438-463b-9026-04bf6b5eb856 to the Glance image server {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 946.289062] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Destroying the VM {{(pid=61906) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 946.289350] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-aa76be7a-3fa8-45a5-a70b-4c15bc03d85a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.297427] env[61906]: DEBUG oslo_vmware.api [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 946.297427] env[61906]: value = "task-1333419" [ 946.297427] env[61906]: _type = "Task" [ 946.297427] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.307064] env[61906]: DEBUG oslo_vmware.api [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333419, 'name': Destroy_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.344331] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "150bf47d-7da9-4e13-ad00-7a8a25b9504e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.344587] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "150bf47d-7da9-4e13-ad00-7a8a25b9504e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.366915] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333418, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.108427} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.367258] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 946.368139] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfb6354c-8961-482f-9944-fd7dc01beb06 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.390808] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] 83b59194-b86f-416a-a7fd-18f75551c2c0/83b59194-b86f-416a-a7fd-18f75551c2c0.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 946.391469] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ffc97900-8bc7-493f-8304-6c699a7a1766 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.413842] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 946.413842] env[61906]: value = "task-1333420" [ 946.413842] env[61906]: _type = "Task" [ 946.413842] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.423591] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333420, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.525835] env[61906]: DEBUG nova.compute.manager [req-c557694f-ac21-4561-a09d-7328d4c4ea23 req-c730ddfc-ffec-4177-a0cc-675240107df3 service nova] [instance: 0f846f95-2573-4450-9918-a34467d73363] Received event network-vif-plugged-12569035-8b00-43d3-bc6d-5a22464fb53d {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 946.526467] env[61906]: DEBUG oslo_concurrency.lockutils [req-c557694f-ac21-4561-a09d-7328d4c4ea23 req-c730ddfc-ffec-4177-a0cc-675240107df3 service nova] Acquiring lock "0f846f95-2573-4450-9918-a34467d73363-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.526731] env[61906]: DEBUG oslo_concurrency.lockutils [req-c557694f-ac21-4561-a09d-7328d4c4ea23 req-c730ddfc-ffec-4177-a0cc-675240107df3 service nova] Lock "0f846f95-2573-4450-9918-a34467d73363-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.527447] env[61906]: DEBUG oslo_concurrency.lockutils [req-c557694f-ac21-4561-a09d-7328d4c4ea23 req-c730ddfc-ffec-4177-a0cc-675240107df3 service nova] Lock "0f846f95-2573-4450-9918-a34467d73363-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.527657] env[61906]: DEBUG nova.compute.manager [req-c557694f-ac21-4561-a09d-7328d4c4ea23 req-c730ddfc-ffec-4177-a0cc-675240107df3 service nova] [instance: 0f846f95-2573-4450-9918-a34467d73363] No waiting events found dispatching network-vif-plugged-12569035-8b00-43d3-bc6d-5a22464fb53d {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 946.527974] env[61906]: WARNING nova.compute.manager [req-c557694f-ac21-4561-a09d-7328d4c4ea23 req-c730ddfc-ffec-4177-a0cc-675240107df3 service nova] [instance: 0f846f95-2573-4450-9918-a34467d73363] Received unexpected event network-vif-plugged-12569035-8b00-43d3-bc6d-5a22464fb53d for instance with vm_state building and task_state spawning. [ 946.538623] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0855be2e-4c0a-4562-a4e3-26755435a6fe tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "b1159533-c970-49d6-af42-b954b20d92fb" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.551970] env[61906]: DEBUG oslo_vmware.api [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333415, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.560145] env[61906]: DEBUG nova.network.neutron [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Successfully updated port: 12569035-8b00-43d3-bc6d-5a22464fb53d {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 946.583981] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.584048] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.584762] env[61906]: DEBUG nova.objects.instance [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61906) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 946.811311] env[61906]: DEBUG oslo_vmware.api [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333419, 'name': Destroy_Task, 'duration_secs': 0.439787} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.811476] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Destroyed the VM [ 946.812044] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Deleting Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 946.812333] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b9ccc9e3-af43-4a71-8256-df1bcb8906d9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.821832] env[61906]: DEBUG oslo_vmware.api [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 946.821832] env[61906]: value = "task-1333421" [ 946.821832] env[61906]: _type = "Task" [ 946.821832] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.830584] env[61906]: DEBUG oslo_vmware.api [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333421, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.847235] env[61906]: DEBUG nova.compute.manager [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 946.925041] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333420, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.050362] env[61906]: DEBUG oslo_vmware.api [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333415, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.063652] env[61906]: DEBUG oslo_concurrency.lockutils [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "refresh_cache-0f846f95-2573-4450-9918-a34467d73363" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.063652] env[61906]: DEBUG oslo_concurrency.lockutils [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired lock "refresh_cache-0f846f95-2573-4450-9918-a34467d73363" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.063652] env[61906]: DEBUG nova.network.neutron [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 947.334024] env[61906]: DEBUG oslo_vmware.api [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333421, 'name': RemoveSnapshot_Task, 'duration_secs': 0.450376} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.334386] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Deleted Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 947.334648] env[61906]: INFO nova.compute.manager [None req-51e75ed1-f0c5-45f0-a697-bddd8e814c6f tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Took 13.52 seconds to snapshot the instance on the hypervisor. [ 947.372738] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.433096] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333420, 'name': ReconfigVM_Task, 'duration_secs': 0.711641} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.433096] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Reconfigured VM instance instance-00000051 to attach disk [datastore2] 83b59194-b86f-416a-a7fd-18f75551c2c0/83b59194-b86f-416a-a7fd-18f75551c2c0.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 947.433096] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9a881564-2c88-41e0-a847-cf2cac76a53b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.438710] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 947.438710] env[61906]: value = "task-1333422" [ 947.438710] env[61906]: _type = "Task" [ 947.438710] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.451130] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333422, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.550852] env[61906]: DEBUG oslo_vmware.api [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333415, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.595821] env[61906]: DEBUG oslo_concurrency.lockutils [None req-86e9ff16-bebc-4806-bd01-64f7aba61b1f tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.598108] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.225s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.599948] env[61906]: INFO nova.compute.claims [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 947.617783] env[61906]: DEBUG nova.network.neutron [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 947.624467] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0855be2e-4c0a-4562-a4e3-26755435a6fe tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "b1159533-c970-49d6-af42-b954b20d92fb" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.624620] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0855be2e-4c0a-4562-a4e3-26755435a6fe tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "b1159533-c970-49d6-af42-b954b20d92fb" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.624842] env[61906]: INFO nova.compute.manager [None req-0855be2e-4c0a-4562-a4e3-26755435a6fe tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Attaching volume 92bbddc7-aa3f-4194-b70a-b524783386e8 to /dev/sdb [ 947.672687] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-008ef328-8156-4127-9e01-17a747317395 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.685063] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9afd338-c2e9-44a1-8869-b21496aaa3c7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.699185] env[61906]: DEBUG nova.virt.block_device [None req-0855be2e-4c0a-4562-a4e3-26755435a6fe tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Updating existing volume attachment record: 83d487d7-3ba2-4676-852f-283ad52fe845 {{(pid=61906) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 947.744259] env[61906]: DEBUG oslo_concurrency.lockutils [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquiring lock "8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.744513] env[61906]: DEBUG oslo_concurrency.lockutils [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Lock "8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.747692] env[61906]: DEBUG oslo_concurrency.lockutils [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquiring lock "8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.747692] env[61906]: DEBUG oslo_concurrency.lockutils [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Lock "8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.747692] env[61906]: DEBUG oslo_concurrency.lockutils [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Lock "8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.754603] env[61906]: INFO nova.compute.manager [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Terminating instance [ 947.756573] env[61906]: DEBUG oslo_concurrency.lockutils [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquiring lock "refresh_cache-8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.756712] env[61906]: DEBUG oslo_concurrency.lockutils [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquired lock "refresh_cache-8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.756875] env[61906]: DEBUG nova.network.neutron [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 947.831592] env[61906]: DEBUG nova.network.neutron [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Updating instance_info_cache with network_info: [{"id": "12569035-8b00-43d3-bc6d-5a22464fb53d", "address": "fa:16:3e:42:29:bb", "network": {"id": "bc9185e9-f26b-489c-84ee-e6db41ac6283", "bridge": "br-int", "label": "tempest-ServersTestJSON-1167747148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53644a4143a24759a9ff2b5e28b84fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12569035-8b", "ovs_interfaceid": "12569035-8b00-43d3-bc6d-5a22464fb53d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.948994] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333422, 'name': Rename_Task, 'duration_secs': 0.194969} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.949333] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 947.949598] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2ac3017d-17c8-4bdc-93ce-5e8816affb23 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.956342] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 947.956342] env[61906]: value = "task-1333426" [ 947.956342] env[61906]: _type = "Task" [ 947.956342] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.969450] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333426, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.987069] env[61906]: DEBUG oslo_concurrency.lockutils [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "ad2db007-e7aa-4c66-9020-ac4acfcff25a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.987536] env[61906]: DEBUG oslo_concurrency.lockutils [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "ad2db007-e7aa-4c66-9020-ac4acfcff25a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.987878] env[61906]: DEBUG oslo_concurrency.lockutils [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "ad2db007-e7aa-4c66-9020-ac4acfcff25a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.988119] env[61906]: DEBUG oslo_concurrency.lockutils [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "ad2db007-e7aa-4c66-9020-ac4acfcff25a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.988293] env[61906]: DEBUG oslo_concurrency.lockutils [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "ad2db007-e7aa-4c66-9020-ac4acfcff25a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.992137] env[61906]: INFO nova.compute.manager [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Terminating instance [ 947.993088] env[61906]: DEBUG nova.compute.manager [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 947.993306] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 947.994570] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9550f668-351e-4bf8-be33-f6b5c6fc48a0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.003254] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 948.003254] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c203622a-5b29-4574-9da3-18e02eb41ecb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.052808] env[61906]: DEBUG oslo_vmware.api [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333415, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.080977] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 948.081324] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 948.081624] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Deleting the datastore file [datastore1] ad2db007-e7aa-4c66-9020-ac4acfcff25a {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 948.081955] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-98bda74b-adcd-469d-b3d8-32b41fb0eae3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.089302] env[61906]: DEBUG oslo_vmware.api [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 948.089302] env[61906]: value = "task-1333428" [ 948.089302] env[61906]: _type = "Task" [ 948.089302] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.097666] env[61906]: DEBUG oslo_vmware.api [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333428, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.287479] env[61906]: DEBUG nova.network.neutron [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 948.334714] env[61906]: DEBUG oslo_concurrency.lockutils [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Releasing lock "refresh_cache-0f846f95-2573-4450-9918-a34467d73363" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.335100] env[61906]: DEBUG nova.compute.manager [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Instance network_info: |[{"id": "12569035-8b00-43d3-bc6d-5a22464fb53d", "address": "fa:16:3e:42:29:bb", "network": {"id": "bc9185e9-f26b-489c-84ee-e6db41ac6283", "bridge": "br-int", "label": "tempest-ServersTestJSON-1167747148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53644a4143a24759a9ff2b5e28b84fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12569035-8b", "ovs_interfaceid": "12569035-8b00-43d3-bc6d-5a22464fb53d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 948.335672] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:42:29:bb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '12569035-8b00-43d3-bc6d-5a22464fb53d', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 948.343273] env[61906]: DEBUG oslo.service.loopingcall [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 948.343793] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0f846f95-2573-4450-9918-a34467d73363] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 948.344054] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4b7dc480-2313-423a-86af-c5c5232b6e75 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.371389] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 948.371389] env[61906]: value = "task-1333429" [ 948.371389] env[61906]: _type = "Task" [ 948.371389] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.381263] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333429, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.393380] env[61906]: DEBUG nova.network.neutron [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.467743] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333426, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.551289] env[61906]: DEBUG oslo_vmware.api [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333415, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.562081] env[61906]: DEBUG nova.compute.manager [req-90c00b68-d92e-4757-b9e1-552134e909fe req-fe295403-35f4-459a-99d8-0adff4387d78 service nova] [instance: 0f846f95-2573-4450-9918-a34467d73363] Received event network-changed-12569035-8b00-43d3-bc6d-5a22464fb53d {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 948.562081] env[61906]: DEBUG nova.compute.manager [req-90c00b68-d92e-4757-b9e1-552134e909fe req-fe295403-35f4-459a-99d8-0adff4387d78 service nova] [instance: 0f846f95-2573-4450-9918-a34467d73363] Refreshing instance network info cache due to event network-changed-12569035-8b00-43d3-bc6d-5a22464fb53d. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 948.562291] env[61906]: DEBUG oslo_concurrency.lockutils [req-90c00b68-d92e-4757-b9e1-552134e909fe req-fe295403-35f4-459a-99d8-0adff4387d78 service nova] Acquiring lock "refresh_cache-0f846f95-2573-4450-9918-a34467d73363" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.562418] env[61906]: DEBUG oslo_concurrency.lockutils [req-90c00b68-d92e-4757-b9e1-552134e909fe req-fe295403-35f4-459a-99d8-0adff4387d78 service nova] Acquired lock "refresh_cache-0f846f95-2573-4450-9918-a34467d73363" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.562620] env[61906]: DEBUG nova.network.neutron [req-90c00b68-d92e-4757-b9e1-552134e909fe req-fe295403-35f4-459a-99d8-0adff4387d78 service nova] [instance: 0f846f95-2573-4450-9918-a34467d73363] Refreshing network info cache for port 12569035-8b00-43d3-bc6d-5a22464fb53d {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 948.600761] env[61906]: DEBUG oslo_vmware.api [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333428, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.183689} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.601048] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 948.601239] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 948.601416] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 948.601840] env[61906]: INFO nova.compute.manager [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Took 0.61 seconds to destroy the instance on the hypervisor. [ 948.601840] env[61906]: DEBUG oslo.service.loopingcall [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 948.602039] env[61906]: DEBUG nova.compute.manager [-] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 948.602127] env[61906]: DEBUG nova.network.neutron [-] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 948.632299] env[61906]: DEBUG oslo_concurrency.lockutils [None req-519db265-c6fa-45ee-bd2b-60fdb5fe0763 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.632630] env[61906]: DEBUG oslo_concurrency.lockutils [None req-519db265-c6fa-45ee-bd2b-60fdb5fe0763 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.632872] env[61906]: DEBUG nova.compute.manager [None req-519db265-c6fa-45ee-bd2b-60fdb5fe0763 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Going to confirm migration 1 {{(pid=61906) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 948.636346] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "7142e219-6150-4c6f-9632-cbf489035431" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.636346] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "7142e219-6150-4c6f-9632-cbf489035431" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.649820] env[61906]: DEBUG oslo_concurrency.lockutils [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "03d3fb1b-0e46-4544-b01d-498a2baf3b45" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.650062] env[61906]: DEBUG oslo_concurrency.lockutils [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "03d3fb1b-0e46-4544-b01d-498a2baf3b45" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.721127] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "bddfdc7f-09d7-4887-b330-34a596ffa562" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.721373] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "bddfdc7f-09d7-4887-b330-34a596ffa562" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.847913] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dbcc9f4-3781-46a8-bc69-02edf92fcef1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.855940] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97716ac4-3b9b-4791-a7a9-c4889ec370f5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.889177] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1bfc888-3859-41b0-b974-471a9d014ab6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.896991] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333429, 'name': CreateVM_Task, 'duration_secs': 0.336154} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.898982] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0f846f95-2573-4450-9918-a34467d73363] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 948.899749] env[61906]: DEBUG oslo_concurrency.lockutils [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.899925] env[61906]: DEBUG oslo_concurrency.lockutils [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.900264] env[61906]: DEBUG oslo_concurrency.lockutils [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 948.901500] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3c022d7-5c37-49fb-b8c9-f332c961d8a6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.906414] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65296c5a-3053-4259-923b-1ec7155b2388 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.908336] env[61906]: DEBUG oslo_concurrency.lockutils [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Releasing lock "refresh_cache-8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.908717] env[61906]: DEBUG nova.compute.manager [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 948.908913] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 948.909919] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcb74096-dec0-4b5f-a12d-dc2a8b0a78a7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.924796] env[61906]: DEBUG nova.compute.provider_tree [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.925973] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 948.925973] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52024628-8de4-98d2-fc73-042a727f6a1d" [ 948.925973] env[61906]: _type = "Task" [ 948.925973] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.928363] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 948.931252] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-17670377-fad7-44cd-89b8-ef4b706eaa4e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.938218] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52024628-8de4-98d2-fc73-042a727f6a1d, 'name': SearchDatastore_Task, 'duration_secs': 0.009456} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.939344] env[61906]: DEBUG oslo_concurrency.lockutils [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.939536] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 948.939747] env[61906]: DEBUG oslo_concurrency.lockutils [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.939888] env[61906]: DEBUG oslo_concurrency.lockutils [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.940078] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 948.941523] env[61906]: DEBUG oslo_vmware.api [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 948.941523] env[61906]: value = "task-1333430" [ 948.941523] env[61906]: _type = "Task" [ 948.941523] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.941523] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8ab4b738-dfaa-430a-bd4b-8ec1a7fc6a13 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.950647] env[61906]: DEBUG oslo_vmware.api [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333430, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.951762] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 948.951944] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 948.952693] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76072653-5c1d-4032-b74f-9b780ff9d18a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.957540] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 948.957540] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52cd3b3e-a025-5763-287f-c8d0c1858a40" [ 948.957540] env[61906]: _type = "Task" [ 948.957540] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.975902] env[61906]: DEBUG oslo_vmware.api [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333426, 'name': PowerOnVM_Task, 'duration_secs': 0.548785} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.976019] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52cd3b3e-a025-5763-287f-c8d0c1858a40, 'name': SearchDatastore_Task, 'duration_secs': 0.008046} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.976273] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 948.976482] env[61906]: INFO nova.compute.manager [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Took 8.46 seconds to spawn the instance on the hypervisor. [ 948.976673] env[61906]: DEBUG nova.compute.manager [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 948.977997] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b3932d5-3032-4ab3-b0ab-4c8b44c76e63 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.980745] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a8f8020-a8a8-4539-8001-93f82dd100d4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.992698] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 948.992698] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52325048-2719-7462-2d95-adabcea38bbe" [ 948.992698] env[61906]: _type = "Task" [ 948.992698] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.002641] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52325048-2719-7462-2d95-adabcea38bbe, 'name': SearchDatastore_Task, 'duration_secs': 0.008243} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.002978] env[61906]: DEBUG oslo_concurrency.lockutils [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.003264] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 0f846f95-2573-4450-9918-a34467d73363/0f846f95-2573-4450-9918-a34467d73363.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 949.003513] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b940e47e-705a-4fa4-8edd-115168f01ac2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.010316] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 949.010316] env[61906]: value = "task-1333431" [ 949.010316] env[61906]: _type = "Task" [ 949.010316] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.018126] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333431, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.051514] env[61906]: DEBUG oslo_vmware.api [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333415, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.143158] env[61906]: DEBUG nova.compute.manager [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 949.154023] env[61906]: DEBUG nova.compute.manager [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 949.203918] env[61906]: DEBUG oslo_concurrency.lockutils [None req-519db265-c6fa-45ee-bd2b-60fdb5fe0763 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "refresh_cache-e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.204513] env[61906]: DEBUG oslo_concurrency.lockutils [None req-519db265-c6fa-45ee-bd2b-60fdb5fe0763 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired lock "refresh_cache-e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.204513] env[61906]: DEBUG nova.network.neutron [None req-519db265-c6fa-45ee-bd2b-60fdb5fe0763 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 949.204513] env[61906]: DEBUG nova.objects.instance [None req-519db265-c6fa-45ee-bd2b-60fdb5fe0763 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lazy-loading 'info_cache' on Instance uuid e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 949.224033] env[61906]: DEBUG nova.compute.manager [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 949.340061] env[61906]: DEBUG nova.network.neutron [req-90c00b68-d92e-4757-b9e1-552134e909fe req-fe295403-35f4-459a-99d8-0adff4387d78 service nova] [instance: 0f846f95-2573-4450-9918-a34467d73363] Updated VIF entry in instance network info cache for port 12569035-8b00-43d3-bc6d-5a22464fb53d. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 949.340571] env[61906]: DEBUG nova.network.neutron [req-90c00b68-d92e-4757-b9e1-552134e909fe req-fe295403-35f4-459a-99d8-0adff4387d78 service nova] [instance: 0f846f95-2573-4450-9918-a34467d73363] Updating instance_info_cache with network_info: [{"id": "12569035-8b00-43d3-bc6d-5a22464fb53d", "address": "fa:16:3e:42:29:bb", "network": {"id": "bc9185e9-f26b-489c-84ee-e6db41ac6283", "bridge": "br-int", "label": "tempest-ServersTestJSON-1167747148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53644a4143a24759a9ff2b5e28b84fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12569035-8b", "ovs_interfaceid": "12569035-8b00-43d3-bc6d-5a22464fb53d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.410061] env[61906]: DEBUG nova.network.neutron [-] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.429785] env[61906]: DEBUG nova.scheduler.client.report [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 949.455807] env[61906]: DEBUG oslo_vmware.api [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333430, 'name': PowerOffVM_Task, 'duration_secs': 0.221784} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.456291] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 949.457062] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 949.457062] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2527360a-3505-4910-b15e-02e89361055f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.483458] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 949.483652] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 949.483846] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Deleting the datastore file [datastore2] 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 949.484144] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-86ec0be4-7f81-46db-b2cf-15097e7750b2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.499542] env[61906]: DEBUG oslo_vmware.api [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 949.499542] env[61906]: value = "task-1333433" [ 949.499542] env[61906]: _type = "Task" [ 949.499542] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.511492] env[61906]: INFO nova.compute.manager [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Took 16.82 seconds to build instance. [ 949.521197] env[61906]: DEBUG oslo_vmware.api [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333433, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.524317] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333431, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.554930] env[61906]: DEBUG oslo_vmware.api [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333415, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.676365] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.678863] env[61906]: DEBUG oslo_concurrency.lockutils [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.747253] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.846961] env[61906]: DEBUG oslo_concurrency.lockutils [req-90c00b68-d92e-4757-b9e1-552134e909fe req-fe295403-35f4-459a-99d8-0adff4387d78 service nova] Releasing lock "refresh_cache-0f846f95-2573-4450-9918-a34467d73363" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.913609] env[61906]: INFO nova.compute.manager [-] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Took 1.31 seconds to deallocate network for instance. [ 949.939291] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.342s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.939803] env[61906]: DEBUG nova.compute.manager [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 949.943394] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.267s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.944717] env[61906]: INFO nova.compute.claims [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 950.012593] env[61906]: DEBUG oslo_vmware.api [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333433, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.12547} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.013067] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.013233] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 950.013417] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 950.013620] env[61906]: INFO nova.compute.manager [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Took 1.10 seconds to destroy the instance on the hypervisor. [ 950.013869] env[61906]: DEBUG oslo.service.loopingcall [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.017210] env[61906]: DEBUG nova.compute.manager [-] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 950.017289] env[61906]: DEBUG nova.network.neutron [-] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 950.019091] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ae965d23-d3cf-45ad-bd4c-1b265ba581a4 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "83b59194-b86f-416a-a7fd-18f75551c2c0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.337s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.024689] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333431, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.035271] env[61906]: DEBUG nova.network.neutron [-] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 950.054220] env[61906]: DEBUG oslo_vmware.api [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333415, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.118209] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e94dfd54-06f8-4d4a-949f-5c9188e4b8fa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.125564] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3931092a-bcd7-466d-8af7-dc3d8cf0cbbc tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Suspending the VM {{(pid=61906) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 950.125828] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-6e22ebc3-effe-47cd-b3e5-3df62b26110b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.132360] env[61906]: DEBUG oslo_vmware.api [None req-3931092a-bcd7-466d-8af7-dc3d8cf0cbbc tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 950.132360] env[61906]: value = "task-1333435" [ 950.132360] env[61906]: _type = "Task" [ 950.132360] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.140911] env[61906]: DEBUG oslo_vmware.api [None req-3931092a-bcd7-466d-8af7-dc3d8cf0cbbc tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333435, 'name': SuspendVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.419722] env[61906]: DEBUG oslo_concurrency.lockutils [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.449403] env[61906]: DEBUG nova.compute.utils [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 950.453541] env[61906]: DEBUG nova.compute.manager [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 950.453874] env[61906]: DEBUG nova.network.neutron [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 950.482875] env[61906]: DEBUG nova.network.neutron [None req-519db265-c6fa-45ee-bd2b-60fdb5fe0763 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Updating instance_info_cache with network_info: [{"id": "b54c27f8-7727-4166-af82-5b6fe0d3519e", "address": "fa:16:3e:ae:b6:e5", "network": {"id": "af039bdd-ffd2-483d-9655-e7df48936290", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-111531540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f7fc2042e6549cabe98838990f35842", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb54c27f8-77", "ovs_interfaceid": "b54c27f8-7727-4166-af82-5b6fe0d3519e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.495055] env[61906]: DEBUG nova.policy [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1b017cdae2af4ae8aa069ae23e744797', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '129cdb06b47346b6895df45e73f7b2dd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 950.524320] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333431, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.538785] env[61906]: DEBUG nova.network.neutron [-] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.558082] env[61906]: DEBUG oslo_vmware.api [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333415, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.592220] env[61906]: DEBUG nova.compute.manager [req-bc8d574d-ea4d-482a-94b3-04693bd6a454 req-6baf00ae-5e58-40e6-91b3-9b65e3662f72 service nova] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Received event network-vif-deleted-d6afef85-5b45-461f-9671-7a187b8022fc {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 950.645471] env[61906]: DEBUG oslo_vmware.api [None req-3931092a-bcd7-466d-8af7-dc3d8cf0cbbc tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333435, 'name': SuspendVM_Task} progress is 12%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.893213] env[61906]: DEBUG nova.network.neutron [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Successfully created port: 5c99ae7c-41d1-48de-a53f-fc8d1a6a4ca2 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 950.954382] env[61906]: DEBUG nova.compute.manager [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 950.985588] env[61906]: DEBUG oslo_concurrency.lockutils [None req-519db265-c6fa-45ee-bd2b-60fdb5fe0763 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Releasing lock "refresh_cache-e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.985863] env[61906]: DEBUG nova.objects.instance [None req-519db265-c6fa-45ee-bd2b-60fdb5fe0763 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lazy-loading 'migration_context' on Instance uuid e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.024042] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333431, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.5173} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.026878] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 0f846f95-2573-4450-9918-a34467d73363/0f846f95-2573-4450-9918-a34467d73363.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 951.027134] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 951.027877] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c89bae45-8ab8-4393-a9c5-0f3a8659cc05 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.036606] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 951.036606] env[61906]: value = "task-1333436" [ 951.036606] env[61906]: _type = "Task" [ 951.036606] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.043314] env[61906]: INFO nova.compute.manager [-] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Took 1.03 seconds to deallocate network for instance. [ 951.061791] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333436, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.068659] env[61906]: DEBUG oslo_vmware.api [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333415, 'name': ReconfigVM_Task, 'duration_secs': 5.895843} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.068952] env[61906]: DEBUG oslo_concurrency.lockutils [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Releasing lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.069188] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Reconfigured VM to detach interface {{(pid=61906) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 951.153879] env[61906]: DEBUG oslo_vmware.api [None req-3931092a-bcd7-466d-8af7-dc3d8cf0cbbc tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333435, 'name': SuspendVM_Task} progress is 58%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.269937] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ea1ac48-b777-4a78-b2a4-37573106dcc6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.279345] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb63a254-c17d-4ef1-9496-8da322c4d71b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.315898] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a77121a-5725-46b5-8f7f-1872a6c21c59 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.323834] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63b5d565-5024-4a4d-be86-3bb661342974 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.340466] env[61906]: DEBUG nova.compute.provider_tree [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 951.489277] env[61906]: DEBUG nova.objects.base [None req-519db265-c6fa-45ee-bd2b-60fdb5fe0763 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 951.490739] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea3b87b0-c4b3-46cb-a97a-cd9a02cf7341 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.510507] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6153aff-5371-4062-af20-d0581617c79f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.516604] env[61906]: DEBUG oslo_vmware.api [None req-519db265-c6fa-45ee-bd2b-60fdb5fe0763 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 951.516604] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52cf91b1-0297-2d58-79b7-e6d7762010b0" [ 951.516604] env[61906]: _type = "Task" [ 951.516604] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.525768] env[61906]: DEBUG oslo_vmware.api [None req-519db265-c6fa-45ee-bd2b-60fdb5fe0763 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52cf91b1-0297-2d58-79b7-e6d7762010b0, 'name': SearchDatastore_Task, 'duration_secs': 0.00627} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.526097] env[61906]: DEBUG oslo_concurrency.lockutils [None req-519db265-c6fa-45ee-bd2b-60fdb5fe0763 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.547746] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333436, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.199866} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.548145] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 951.549053] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-658238e4-aa97-401d-bf88-f5589f328afc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.563383] env[61906]: DEBUG oslo_concurrency.lockutils [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.572171] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] 0f846f95-2573-4450-9918-a34467d73363/0f846f95-2573-4450-9918-a34467d73363.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 951.572509] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c71d150-5e10-4a0c-b2b8-43572a5dba0a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.595165] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 951.595165] env[61906]: value = "task-1333437" [ 951.595165] env[61906]: _type = "Task" [ 951.595165] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.605305] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333437, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.644339] env[61906]: DEBUG oslo_vmware.api [None req-3931092a-bcd7-466d-8af7-dc3d8cf0cbbc tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333435, 'name': SuspendVM_Task, 'duration_secs': 1.217713} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.644622] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3931092a-bcd7-466d-8af7-dc3d8cf0cbbc tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Suspended the VM {{(pid=61906) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 951.644818] env[61906]: DEBUG nova.compute.manager [None req-3931092a-bcd7-466d-8af7-dc3d8cf0cbbc tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 951.645615] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-311b628a-4269-4f12-93a9-5cca61144685 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.843439] env[61906]: DEBUG nova.scheduler.client.report [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 951.969820] env[61906]: DEBUG nova.compute.manager [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 951.995444] env[61906]: DEBUG nova.virt.hardware [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 951.995709] env[61906]: DEBUG nova.virt.hardware [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 951.995874] env[61906]: DEBUG nova.virt.hardware [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 951.996154] env[61906]: DEBUG nova.virt.hardware [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 951.996416] env[61906]: DEBUG nova.virt.hardware [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 951.996679] env[61906]: DEBUG nova.virt.hardware [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 951.996931] env[61906]: DEBUG nova.virt.hardware [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 951.997130] env[61906]: DEBUG nova.virt.hardware [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 951.997353] env[61906]: DEBUG nova.virt.hardware [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 951.997565] env[61906]: DEBUG nova.virt.hardware [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 951.997760] env[61906]: DEBUG nova.virt.hardware [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 951.998777] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d20bd53a-39f0-4877-8181-adc1b2388770 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.011464] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ddbbb62-5211-4629-8d7b-faf6b87a61cf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.104992] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333437, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.253494] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-0855be2e-4c0a-4562-a4e3-26755435a6fe tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Volume attach. Driver type: vmdk {{(pid=61906) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 952.254198] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-0855be2e-4c0a-4562-a4e3-26755435a6fe tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284803', 'volume_id': '92bbddc7-aa3f-4194-b70a-b524783386e8', 'name': 'volume-92bbddc7-aa3f-4194-b70a-b524783386e8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b1159533-c970-49d6-af42-b954b20d92fb', 'attached_at': '', 'detached_at': '', 'volume_id': '92bbddc7-aa3f-4194-b70a-b524783386e8', 'serial': '92bbddc7-aa3f-4194-b70a-b524783386e8'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 952.254864] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7082173f-f9bd-4850-8ed3-aa6a371c60f2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.271602] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dec6c49-5421-4b21-b7d2-1120fecbd899 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.296188] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-0855be2e-4c0a-4562-a4e3-26755435a6fe tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] volume-92bbddc7-aa3f-4194-b70a-b524783386e8/volume-92bbddc7-aa3f-4194-b70a-b524783386e8.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 952.296523] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b8fa9f1-8d12-4108-a389-f433a49d479f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.315697] env[61906]: DEBUG oslo_vmware.api [None req-0855be2e-4c0a-4562-a4e3-26755435a6fe tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 952.315697] env[61906]: value = "task-1333438" [ 952.315697] env[61906]: _type = "Task" [ 952.315697] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.326656] env[61906]: DEBUG oslo_vmware.api [None req-0855be2e-4c0a-4562-a4e3-26755435a6fe tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333438, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.335356] env[61906]: DEBUG oslo_concurrency.lockutils [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "refresh_cache-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.335586] env[61906]: DEBUG oslo_concurrency.lockutils [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired lock "refresh_cache-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.335769] env[61906]: DEBUG nova.network.neutron [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 952.348200] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.405s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.352020] env[61906]: DEBUG nova.compute.manager [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 952.352020] env[61906]: DEBUG oslo_concurrency.lockutils [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.673s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.353121] env[61906]: INFO nova.compute.claims [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 952.403701] env[61906]: DEBUG nova.compute.manager [req-edff25e4-913a-41a4-b221-9c413f8e9db9 req-ac492fdc-7c8f-44f0-8f97-e833db39390b service nova] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Received event network-vif-plugged-5c99ae7c-41d1-48de-a53f-fc8d1a6a4ca2 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 952.403964] env[61906]: DEBUG oslo_concurrency.lockutils [req-edff25e4-913a-41a4-b221-9c413f8e9db9 req-ac492fdc-7c8f-44f0-8f97-e833db39390b service nova] Acquiring lock "150bf47d-7da9-4e13-ad00-7a8a25b9504e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.404214] env[61906]: DEBUG oslo_concurrency.lockutils [req-edff25e4-913a-41a4-b221-9c413f8e9db9 req-ac492fdc-7c8f-44f0-8f97-e833db39390b service nova] Lock "150bf47d-7da9-4e13-ad00-7a8a25b9504e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.404422] env[61906]: DEBUG oslo_concurrency.lockutils [req-edff25e4-913a-41a4-b221-9c413f8e9db9 req-ac492fdc-7c8f-44f0-8f97-e833db39390b service nova] Lock "150bf47d-7da9-4e13-ad00-7a8a25b9504e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.404624] env[61906]: DEBUG nova.compute.manager [req-edff25e4-913a-41a4-b221-9c413f8e9db9 req-ac492fdc-7c8f-44f0-8f97-e833db39390b service nova] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] No waiting events found dispatching network-vif-plugged-5c99ae7c-41d1-48de-a53f-fc8d1a6a4ca2 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 952.404823] env[61906]: WARNING nova.compute.manager [req-edff25e4-913a-41a4-b221-9c413f8e9db9 req-ac492fdc-7c8f-44f0-8f97-e833db39390b service nova] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Received unexpected event network-vif-plugged-5c99ae7c-41d1-48de-a53f-fc8d1a6a4ca2 for instance with vm_state building and task_state spawning. [ 952.544436] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "83b59194-b86f-416a-a7fd-18f75551c2c0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.544723] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "83b59194-b86f-416a-a7fd-18f75551c2c0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.545015] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "83b59194-b86f-416a-a7fd-18f75551c2c0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.545167] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "83b59194-b86f-416a-a7fd-18f75551c2c0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.545375] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "83b59194-b86f-416a-a7fd-18f75551c2c0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.547999] env[61906]: INFO nova.compute.manager [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Terminating instance [ 952.550037] env[61906]: DEBUG nova.compute.manager [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 952.550255] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 952.551101] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acaeab01-29fb-4079-a660-5e64a3c4b973 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.558598] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 952.558844] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-14313286-bb4e-41c4-a46c-3c095b9f2f6e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.607203] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333437, 'name': ReconfigVM_Task, 'duration_secs': 0.862937} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.607584] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Reconfigured VM instance instance-00000052 to attach disk [datastore2] 0f846f95-2573-4450-9918-a34467d73363/0f846f95-2573-4450-9918-a34467d73363.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 952.609037] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-569531ef-97db-46cd-aa8d-c5a6f3d4e0ae {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.616417] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 952.616417] env[61906]: value = "task-1333440" [ 952.616417] env[61906]: _type = "Task" [ 952.616417] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.622075] env[61906]: DEBUG nova.compute.manager [req-363ee9e3-0e04-40ba-829c-f5c682030bc6 req-3c7455fd-92a6-49e8-b1d8-086ed93c471d service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Received event network-vif-deleted-c2232d91-1c2c-4a66-9026-933169da90e1 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 952.622308] env[61906]: INFO nova.compute.manager [req-363ee9e3-0e04-40ba-829c-f5c682030bc6 req-3c7455fd-92a6-49e8-b1d8-086ed93c471d service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Neutron deleted interface c2232d91-1c2c-4a66-9026-933169da90e1; detaching it from the instance and deleting it from the info cache [ 952.622678] env[61906]: DEBUG nova.network.neutron [req-363ee9e3-0e04-40ba-829c-f5c682030bc6 req-3c7455fd-92a6-49e8-b1d8-086ed93c471d service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Updating instance_info_cache with network_info: [{"id": "3e04c5c7-8d7a-46ae-a2d7-52dc429650f9", "address": "fa:16:3e:f1:15:76", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e04c5c7-8d", "ovs_interfaceid": "3e04c5c7-8d7a-46ae-a2d7-52dc429650f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "4f8ab1c1-fd91-48d3-90e0-ed5ff9fa0b4d", "address": "fa:16:3e:ee:f2:da", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f8ab1c1-fd", "ovs_interfaceid": "4f8ab1c1-fd91-48d3-90e0-ed5ff9fa0b4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.630543] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333440, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.631402] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 952.631605] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 952.631787] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Deleting the datastore file [datastore2] 83b59194-b86f-416a-a7fd-18f75551c2c0 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 952.632050] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6beaaf68-c93d-4195-80d3-0e4f9aed3dee {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.636870] env[61906]: DEBUG oslo_concurrency.lockutils [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.637123] env[61906]: DEBUG oslo_concurrency.lockutils [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.637332] env[61906]: DEBUG oslo_concurrency.lockutils [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.637522] env[61906]: DEBUG oslo_concurrency.lockutils [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.637691] env[61906]: DEBUG oslo_concurrency.lockutils [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.640964] env[61906]: DEBUG oslo_vmware.api [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 952.640964] env[61906]: value = "task-1333441" [ 952.640964] env[61906]: _type = "Task" [ 952.640964] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.641399] env[61906]: INFO nova.compute.manager [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Terminating instance [ 952.646637] env[61906]: DEBUG nova.compute.manager [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 952.646837] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 952.647788] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ba9f1c-cfd1-435e-9190-713e36995282 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.655046] env[61906]: DEBUG oslo_vmware.api [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333441, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.657122] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 952.657404] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d8ce5582-77f4-43f6-b2d7-b8c6816a0654 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.662720] env[61906]: DEBUG oslo_vmware.api [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 952.662720] env[61906]: value = "task-1333442" [ 952.662720] env[61906]: _type = "Task" [ 952.662720] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.670899] env[61906]: DEBUG oslo_vmware.api [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333442, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.826679] env[61906]: DEBUG oslo_vmware.api [None req-0855be2e-4c0a-4562-a4e3-26755435a6fe tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333438, 'name': ReconfigVM_Task, 'duration_secs': 0.347714} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.827012] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-0855be2e-4c0a-4562-a4e3-26755435a6fe tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Reconfigured VM instance instance-00000048 to attach disk [datastore2] volume-92bbddc7-aa3f-4194-b70a-b524783386e8/volume-92bbddc7-aa3f-4194-b70a-b524783386e8.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 952.831825] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-77f2c828-0ab5-49b9-8bd1-6832ae242aca {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.848943] env[61906]: DEBUG oslo_vmware.api [None req-0855be2e-4c0a-4562-a4e3-26755435a6fe tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 952.848943] env[61906]: value = "task-1333443" [ 952.848943] env[61906]: _type = "Task" [ 952.848943] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.857669] env[61906]: DEBUG nova.compute.utils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 952.859024] env[61906]: DEBUG oslo_vmware.api [None req-0855be2e-4c0a-4562-a4e3-26755435a6fe tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333443, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.861217] env[61906]: DEBUG nova.compute.manager [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 952.861300] env[61906]: DEBUG nova.network.neutron [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 952.932367] env[61906]: DEBUG nova.policy [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '23ff59c262484422a8badcbca5a103f9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '76ca0b5a4fd24ede8eac4851fc365848', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 952.966674] env[61906]: DEBUG nova.network.neutron [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Successfully updated port: 5c99ae7c-41d1-48de-a53f-fc8d1a6a4ca2 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 952.997707] env[61906]: DEBUG nova.compute.manager [req-94bc0328-a1e8-4f31-be81-17f266a1830d req-bd7070a3-0dc2-44f5-bcc7-9798175a2cd9 service nova] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Received event network-changed-5c99ae7c-41d1-48de-a53f-fc8d1a6a4ca2 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 952.997889] env[61906]: DEBUG nova.compute.manager [req-94bc0328-a1e8-4f31-be81-17f266a1830d req-bd7070a3-0dc2-44f5-bcc7-9798175a2cd9 service nova] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Refreshing instance network info cache due to event network-changed-5c99ae7c-41d1-48de-a53f-fc8d1a6a4ca2. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 952.998115] env[61906]: DEBUG oslo_concurrency.lockutils [req-94bc0328-a1e8-4f31-be81-17f266a1830d req-bd7070a3-0dc2-44f5-bcc7-9798175a2cd9 service nova] Acquiring lock "refresh_cache-150bf47d-7da9-4e13-ad00-7a8a25b9504e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.998217] env[61906]: DEBUG oslo_concurrency.lockutils [req-94bc0328-a1e8-4f31-be81-17f266a1830d req-bd7070a3-0dc2-44f5-bcc7-9798175a2cd9 service nova] Acquired lock "refresh_cache-150bf47d-7da9-4e13-ad00-7a8a25b9504e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.998388] env[61906]: DEBUG nova.network.neutron [req-94bc0328-a1e8-4f31-be81-17f266a1830d req-bd7070a3-0dc2-44f5-bcc7-9798175a2cd9 service nova] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Refreshing network info cache for port 5c99ae7c-41d1-48de-a53f-fc8d1a6a4ca2 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 953.129273] env[61906]: DEBUG oslo_concurrency.lockutils [req-363ee9e3-0e04-40ba-829c-f5c682030bc6 req-3c7455fd-92a6-49e8-b1d8-086ed93c471d service nova] Acquiring lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.129733] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333440, 'name': Rename_Task, 'duration_secs': 0.178331} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.130119] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 953.130511] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0cf6bf06-df5a-45ee-924a-dc683679c5a7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.138840] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 953.138840] env[61906]: value = "task-1333444" [ 953.138840] env[61906]: _type = "Task" [ 953.138840] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.153241] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333444, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.156790] env[61906]: DEBUG oslo_vmware.api [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333441, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154332} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.157127] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 953.157265] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 953.157448] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 953.157627] env[61906]: INFO nova.compute.manager [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Took 0.61 seconds to destroy the instance on the hypervisor. [ 953.157872] env[61906]: DEBUG oslo.service.loopingcall [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 953.158079] env[61906]: DEBUG nova.compute.manager [-] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 953.158175] env[61906]: DEBUG nova.network.neutron [-] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 953.171338] env[61906]: DEBUG oslo_vmware.api [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333442, 'name': PowerOffVM_Task, 'duration_secs': 0.188004} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.171611] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 953.171783] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 953.172103] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-16642310-063b-4722-b017-ec86e51c9b9e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.241608] env[61906]: INFO nova.network.neutron [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Port c2232d91-1c2c-4a66-9026-933169da90e1 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 953.241864] env[61906]: INFO nova.network.neutron [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Port 4f8ab1c1-fd91-48d3-90e0-ed5ff9fa0b4d from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 953.242323] env[61906]: DEBUG nova.network.neutron [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Updating instance_info_cache with network_info: [{"id": "3e04c5c7-8d7a-46ae-a2d7-52dc429650f9", "address": "fa:16:3e:f1:15:76", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e04c5c7-8d", "ovs_interfaceid": "3e04c5c7-8d7a-46ae-a2d7-52dc429650f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.280807] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 953.281123] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 953.281412] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Deleting the datastore file [datastore2] a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 953.281728] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b26e0675-b0af-4efb-b3ae-bdf5192cc543 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.288062] env[61906]: DEBUG oslo_vmware.api [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 953.288062] env[61906]: value = "task-1333446" [ 953.288062] env[61906]: _type = "Task" [ 953.288062] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.297150] env[61906]: DEBUG oslo_vmware.api [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333446, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.359237] env[61906]: DEBUG oslo_vmware.api [None req-0855be2e-4c0a-4562-a4e3-26755435a6fe tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333443, 'name': ReconfigVM_Task, 'duration_secs': 0.144276} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.362485] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-0855be2e-4c0a-4562-a4e3-26755435a6fe tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284803', 'volume_id': '92bbddc7-aa3f-4194-b70a-b524783386e8', 'name': 'volume-92bbddc7-aa3f-4194-b70a-b524783386e8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b1159533-c970-49d6-af42-b954b20d92fb', 'attached_at': '', 'detached_at': '', 'volume_id': '92bbddc7-aa3f-4194-b70a-b524783386e8', 'serial': '92bbddc7-aa3f-4194-b70a-b524783386e8'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 953.364037] env[61906]: DEBUG nova.compute.manager [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 953.368027] env[61906]: DEBUG nova.network.neutron [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Successfully created port: ff7dc3ad-711f-4510-8a80-43f9717aeeba {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 953.466601] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "refresh_cache-150bf47d-7da9-4e13-ad00-7a8a25b9504e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.539493] env[61906]: DEBUG nova.network.neutron [req-94bc0328-a1e8-4f31-be81-17f266a1830d req-bd7070a3-0dc2-44f5-bcc7-9798175a2cd9 service nova] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 953.598927] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f57cc8f-844c-4e4b-8e7b-1719728491af {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.607125] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e3aa3e6-29a4-41de-b08b-c405cf919406 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.613800] env[61906]: DEBUG nova.network.neutron [req-94bc0328-a1e8-4f31-be81-17f266a1830d req-bd7070a3-0dc2-44f5-bcc7-9798175a2cd9 service nova] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.642712] env[61906]: DEBUG oslo_concurrency.lockutils [req-94bc0328-a1e8-4f31-be81-17f266a1830d req-bd7070a3-0dc2-44f5-bcc7-9798175a2cd9 service nova] Releasing lock "refresh_cache-150bf47d-7da9-4e13-ad00-7a8a25b9504e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.648037] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d753969-10b5-464d-b675-9b08b512cd6d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.648944] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquired lock "refresh_cache-150bf47d-7da9-4e13-ad00-7a8a25b9504e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.649118] env[61906]: DEBUG nova.network.neutron [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 953.659309] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8d98f42-720b-48ee-b0a3-6d9041343181 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.663230] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333444, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.673820] env[61906]: DEBUG nova.compute.provider_tree [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 953.746648] env[61906]: DEBUG oslo_concurrency.lockutils [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Releasing lock "refresh_cache-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.800236] env[61906]: DEBUG oslo_vmware.api [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333446, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165526} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.800416] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 953.800662] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 953.800885] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 953.801162] env[61906]: INFO nova.compute.manager [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Took 1.15 seconds to destroy the instance on the hypervisor. [ 953.801451] env[61906]: DEBUG oslo.service.loopingcall [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 953.801732] env[61906]: DEBUG nova.compute.manager [-] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 953.801897] env[61906]: DEBUG nova.network.neutron [-] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 953.865118] env[61906]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 4f8ab1c1-fd91-48d3-90e0-ed5ff9fa0b4d could not be found.", "detail": ""}} {{(pid=61906) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 953.865342] env[61906]: DEBUG nova.network.neutron [-] Unable to show port 4f8ab1c1-fd91-48d3-90e0-ed5ff9fa0b4d as it no longer exists. {{(pid=61906) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 953.956036] env[61906]: DEBUG nova.network.neutron [-] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.153891] env[61906]: DEBUG oslo_vmware.api [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333444, 'name': PowerOnVM_Task, 'duration_secs': 0.526075} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.154195] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 954.154405] env[61906]: INFO nova.compute.manager [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Took 8.53 seconds to spawn the instance on the hypervisor. [ 954.154582] env[61906]: DEBUG nova.compute.manager [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 954.155838] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b43616d0-6a33-449c-a873-db15454fc583 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.176835] env[61906]: DEBUG nova.scheduler.client.report [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 954.201101] env[61906]: DEBUG nova.network.neutron [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 954.250722] env[61906]: DEBUG oslo_concurrency.lockutils [None req-65082d2d-21e4-4493-8873-570b2d095890 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "interface-a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8-c2232d91-1c2c-4a66-9026-933169da90e1" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.805s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.372797] env[61906]: DEBUG nova.network.neutron [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Updating instance_info_cache with network_info: [{"id": "5c99ae7c-41d1-48de-a53f-fc8d1a6a4ca2", "address": "fa:16:3e:b2:fc:be", "network": {"id": "c3bcef0f-44a7-44f7-8130-df77c0d3ac2d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-330384233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "129cdb06b47346b6895df45e73f7b2dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35e463c7-7d78-4d66-8efd-6127b1f3ee17", "external-id": "nsx-vlan-transportzone-175", "segmentation_id": 175, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c99ae7c-41", "ovs_interfaceid": "5c99ae7c-41d1-48de-a53f-fc8d1a6a4ca2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.381667] env[61906]: DEBUG nova.compute.manager [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 954.397398] env[61906]: DEBUG nova.objects.instance [None req-0855be2e-4c0a-4562-a4e3-26755435a6fe tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lazy-loading 'flavor' on Instance uuid b1159533-c970-49d6-af42-b954b20d92fb {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.415176] env[61906]: DEBUG nova.virt.hardware [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 954.415176] env[61906]: DEBUG nova.virt.hardware [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 954.415176] env[61906]: DEBUG nova.virt.hardware [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 954.415433] env[61906]: DEBUG nova.virt.hardware [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 954.415634] env[61906]: DEBUG nova.virt.hardware [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 954.415829] env[61906]: DEBUG nova.virt.hardware [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 954.416101] env[61906]: DEBUG nova.virt.hardware [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 954.416308] env[61906]: DEBUG nova.virt.hardware [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 954.417204] env[61906]: DEBUG nova.virt.hardware [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 954.417204] env[61906]: DEBUG nova.virt.hardware [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 954.417204] env[61906]: DEBUG nova.virt.hardware [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 954.417856] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9466f74f-c0f6-4b46-923d-cb8060e0def8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.428162] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f6537b3-76d0-4f3b-8e9b-3915167d2f6f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.461544] env[61906]: INFO nova.compute.manager [-] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Took 1.30 seconds to deallocate network for instance. [ 954.676924] env[61906]: INFO nova.compute.manager [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Took 15.67 seconds to build instance. [ 954.680873] env[61906]: DEBUG oslo_concurrency.lockutils [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.329s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.681400] env[61906]: DEBUG nova.compute.manager [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 954.684776] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.938s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.685247] env[61906]: INFO nova.compute.claims [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 954.875837] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Releasing lock "refresh_cache-150bf47d-7da9-4e13-ad00-7a8a25b9504e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.876311] env[61906]: DEBUG nova.compute.manager [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Instance network_info: |[{"id": "5c99ae7c-41d1-48de-a53f-fc8d1a6a4ca2", "address": "fa:16:3e:b2:fc:be", "network": {"id": "c3bcef0f-44a7-44f7-8130-df77c0d3ac2d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-330384233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "129cdb06b47346b6895df45e73f7b2dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35e463c7-7d78-4d66-8efd-6127b1f3ee17", "external-id": "nsx-vlan-transportzone-175", "segmentation_id": 175, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c99ae7c-41", "ovs_interfaceid": "5c99ae7c-41d1-48de-a53f-fc8d1a6a4ca2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 954.877185] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b2:fc:be', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35e463c7-7d78-4d66-8efd-6127b1f3ee17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5c99ae7c-41d1-48de-a53f-fc8d1a6a4ca2', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 954.885253] env[61906]: DEBUG oslo.service.loopingcall [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 954.885528] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 954.885796] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ab45d01d-6eba-4910-96db-302b70627fbd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.903192] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0855be2e-4c0a-4562-a4e3-26755435a6fe tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "b1159533-c970-49d6-af42-b954b20d92fb" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.279s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.910651] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 954.910651] env[61906]: value = "task-1333447" [ 954.910651] env[61906]: _type = "Task" [ 954.910651] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.916904] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333447, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.968644] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.000657] env[61906]: DEBUG nova.network.neutron [-] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.078016] env[61906]: DEBUG nova.compute.manager [req-9f703b5c-fa2f-4df8-9bcf-9ce077433311 req-7841fb67-d4e9-4f89-89aa-df5ff3ba9ca2 service nova] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Received event network-vif-plugged-ff7dc3ad-711f-4510-8a80-43f9717aeeba {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.078016] env[61906]: DEBUG oslo_concurrency.lockutils [req-9f703b5c-fa2f-4df8-9bcf-9ce077433311 req-7841fb67-d4e9-4f89-89aa-df5ff3ba9ca2 service nova] Acquiring lock "7142e219-6150-4c6f-9632-cbf489035431-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.078016] env[61906]: DEBUG oslo_concurrency.lockutils [req-9f703b5c-fa2f-4df8-9bcf-9ce077433311 req-7841fb67-d4e9-4f89-89aa-df5ff3ba9ca2 service nova] Lock "7142e219-6150-4c6f-9632-cbf489035431-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.078016] env[61906]: DEBUG oslo_concurrency.lockutils [req-9f703b5c-fa2f-4df8-9bcf-9ce077433311 req-7841fb67-d4e9-4f89-89aa-df5ff3ba9ca2 service nova] Lock "7142e219-6150-4c6f-9632-cbf489035431-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.078016] env[61906]: DEBUG nova.compute.manager [req-9f703b5c-fa2f-4df8-9bcf-9ce077433311 req-7841fb67-d4e9-4f89-89aa-df5ff3ba9ca2 service nova] [instance: 7142e219-6150-4c6f-9632-cbf489035431] No waiting events found dispatching network-vif-plugged-ff7dc3ad-711f-4510-8a80-43f9717aeeba {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 955.078016] env[61906]: WARNING nova.compute.manager [req-9f703b5c-fa2f-4df8-9bcf-9ce077433311 req-7841fb67-d4e9-4f89-89aa-df5ff3ba9ca2 service nova] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Received unexpected event network-vif-plugged-ff7dc3ad-711f-4510-8a80-43f9717aeeba for instance with vm_state building and task_state spawning. [ 955.105943] env[61906]: DEBUG nova.compute.manager [req-e27608ed-788f-47c4-a6bc-781f53968c95 req-b02a9fd3-64af-4129-805d-d56782c14c1f service nova] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Received event network-vif-deleted-fb758a9b-f65f-4d8c-84a3-290128c68110 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.105943] env[61906]: DEBUG nova.compute.manager [req-e27608ed-788f-47c4-a6bc-781f53968c95 req-b02a9fd3-64af-4129-805d-d56782c14c1f service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Received event network-vif-deleted-3e04c5c7-8d7a-46ae-a2d7-52dc429650f9 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.111180] env[61906]: DEBUG oslo_concurrency.lockutils [None req-19a38656-c3fc-444c-a9e7-303bee0d79cb tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "b1159533-c970-49d6-af42-b954b20d92fb" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.111452] env[61906]: DEBUG oslo_concurrency.lockutils [None req-19a38656-c3fc-444c-a9e7-303bee0d79cb tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "b1159533-c970-49d6-af42-b954b20d92fb" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.002s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.111640] env[61906]: DEBUG nova.compute.manager [None req-19a38656-c3fc-444c-a9e7-303bee0d79cb tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 955.112607] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-482a1591-f89e-4ca6-a652-d839d9012046 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.120541] env[61906]: DEBUG nova.compute.manager [None req-19a38656-c3fc-444c-a9e7-303bee0d79cb tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61906) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 955.121501] env[61906]: DEBUG nova.objects.instance [None req-19a38656-c3fc-444c-a9e7-303bee0d79cb tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lazy-loading 'flavor' on Instance uuid b1159533-c970-49d6-af42-b954b20d92fb {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 955.132615] env[61906]: DEBUG nova.network.neutron [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Successfully updated port: ff7dc3ad-711f-4510-8a80-43f9717aeeba {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 955.178682] env[61906]: DEBUG oslo_concurrency.lockutils [None req-804559bc-faee-41a5-98a5-587c45d1aa55 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "0f846f95-2573-4450-9918-a34467d73363" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.177s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.189149] env[61906]: DEBUG nova.compute.utils [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 955.193638] env[61906]: DEBUG nova.compute.manager [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 955.193638] env[61906]: DEBUG nova.network.neutron [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 955.231734] env[61906]: DEBUG nova.policy [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc6b374325bc4ef9882e5f84ddd62fd1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f2d2198383d74dac90aa9727a62176df', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 955.420509] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333447, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.472428] env[61906]: DEBUG nova.network.neutron [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Successfully created port: b5eb6e7b-4b1c-411f-8d99-ad4888193eda {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 955.507664] env[61906]: INFO nova.compute.manager [-] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Took 1.71 seconds to deallocate network for instance. [ 955.627398] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-19a38656-c3fc-444c-a9e7-303bee0d79cb tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 955.627788] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7266fcfc-5deb-4b68-964d-fac4607ce4da {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.634446] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "refresh_cache-7142e219-6150-4c6f-9632-cbf489035431" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.634586] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquired lock "refresh_cache-7142e219-6150-4c6f-9632-cbf489035431" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.634730] env[61906]: DEBUG nova.network.neutron [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 955.636674] env[61906]: DEBUG oslo_vmware.api [None req-19a38656-c3fc-444c-a9e7-303bee0d79cb tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 955.636674] env[61906]: value = "task-1333448" [ 955.636674] env[61906]: _type = "Task" [ 955.636674] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.645196] env[61906]: DEBUG oslo_vmware.api [None req-19a38656-c3fc-444c-a9e7-303bee0d79cb tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333448, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.694007] env[61906]: DEBUG nova.compute.manager [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 955.702320] env[61906]: DEBUG oslo_concurrency.lockutils [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "f49fb403-fbb4-4e26-8e75-c160d11dea05" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.702619] env[61906]: DEBUG oslo_concurrency.lockutils [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "f49fb403-fbb4-4e26-8e75-c160d11dea05" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.974955] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333447, 'name': CreateVM_Task, 'duration_secs': 0.58118} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.974955] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 955.974955] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.974955] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.974955] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 955.974955] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d76099e9-a79d-4686-91b4-78368014062d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.974955] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 955.974955] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5224caa0-03ff-b6f6-ae62-06fa54e2fbed" [ 955.974955] env[61906]: _type = "Task" [ 955.974955] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.974955] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5224caa0-03ff-b6f6-ae62-06fa54e2fbed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.974955] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c7a8d6b-6141-443b-b977-681ae8e23d80 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.980456] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ac5e0b1-6f66-48b8-b114-7fa8577fec81 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.009925] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83e5c3e9-d727-4efb-9226-0492398538bc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.016862] env[61906]: DEBUG oslo_concurrency.lockutils [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.018070] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59d66730-d910-4a8d-8ec7-eee23a8ad278 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.038802] env[61906]: DEBUG nova.compute.provider_tree [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.148152] env[61906]: DEBUG oslo_vmware.api [None req-19a38656-c3fc-444c-a9e7-303bee0d79cb tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333448, 'name': PowerOffVM_Task, 'duration_secs': 0.180077} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.148447] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-19a38656-c3fc-444c-a9e7-303bee0d79cb tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 956.149122] env[61906]: DEBUG nova.compute.manager [None req-19a38656-c3fc-444c-a9e7-303bee0d79cb tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 956.149397] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302558ac-4a06-454a-9869-38d64099acbe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.169812] env[61906]: DEBUG nova.network.neutron [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 956.205592] env[61906]: DEBUG nova.compute.manager [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 956.313493] env[61906]: DEBUG nova.network.neutron [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Updating instance_info_cache with network_info: [{"id": "ff7dc3ad-711f-4510-8a80-43f9717aeeba", "address": "fa:16:3e:ef:52:74", "network": {"id": "0a3274f7-785f-45e2-9582-0b611955b024", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-593827400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76ca0b5a4fd24ede8eac4851fc365848", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff7dc3ad-71", "ovs_interfaceid": "ff7dc3ad-711f-4510-8a80-43f9717aeeba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.459068] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5224caa0-03ff-b6f6-ae62-06fa54e2fbed, 'name': SearchDatastore_Task, 'duration_secs': 0.009566} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.459262] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.459513] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 956.459769] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.459926] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.460128] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 956.460417] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-27bff068-44d2-4d0b-88c3-6db5522f03b9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.469018] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 956.469246] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 956.469979] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db34c485-acf0-4d9f-ad42-9c635c75d9ac {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.475031] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 956.475031] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52bb2e3b-f020-447b-9a0a-0559903d54b4" [ 956.475031] env[61906]: _type = "Task" [ 956.475031] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.482614] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52bb2e3b-f020-447b-9a0a-0559903d54b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.542164] env[61906]: DEBUG nova.scheduler.client.report [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 956.662132] env[61906]: DEBUG oslo_concurrency.lockutils [None req-19a38656-c3fc-444c-a9e7-303bee0d79cb tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "b1159533-c970-49d6-af42-b954b20d92fb" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.551s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.705190] env[61906]: DEBUG nova.compute.manager [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 956.722420] env[61906]: DEBUG oslo_concurrency.lockutils [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.729916] env[61906]: DEBUG nova.virt.hardware [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 956.730186] env[61906]: DEBUG nova.virt.hardware [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 956.730348] env[61906]: DEBUG nova.virt.hardware [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 956.730582] env[61906]: DEBUG nova.virt.hardware [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 956.730736] env[61906]: DEBUG nova.virt.hardware [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 956.730889] env[61906]: DEBUG nova.virt.hardware [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 956.731104] env[61906]: DEBUG nova.virt.hardware [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 956.731269] env[61906]: DEBUG nova.virt.hardware [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 956.731439] env[61906]: DEBUG nova.virt.hardware [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 956.731606] env[61906]: DEBUG nova.virt.hardware [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 956.731815] env[61906]: DEBUG nova.virt.hardware [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 956.733030] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2e1cd85-bc04-4740-954f-cbbf6e4f146e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.741739] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb5800be-f467-4bec-a5a0-af89e9904ce3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.816556] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Releasing lock "refresh_cache-7142e219-6150-4c6f-9632-cbf489035431" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.818052] env[61906]: DEBUG nova.compute.manager [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Instance network_info: |[{"id": "ff7dc3ad-711f-4510-8a80-43f9717aeeba", "address": "fa:16:3e:ef:52:74", "network": {"id": "0a3274f7-785f-45e2-9582-0b611955b024", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-593827400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76ca0b5a4fd24ede8eac4851fc365848", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff7dc3ad-71", "ovs_interfaceid": "ff7dc3ad-711f-4510-8a80-43f9717aeeba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 956.818052] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ef:52:74', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ff7dc3ad-711f-4510-8a80-43f9717aeeba', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 956.825298] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Creating folder: Project (76ca0b5a4fd24ede8eac4851fc365848). Parent ref: group-v284713. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 956.825566] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a4170512-683e-4c0e-95a6-e9d63b4c3a12 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.835831] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Created folder: Project (76ca0b5a4fd24ede8eac4851fc365848) in parent group-v284713. [ 956.836040] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Creating folder: Instances. Parent ref: group-v284806. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 956.836287] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e9f30cc1-bc32-475c-8e14-6bde0193f90a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.845373] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Created folder: Instances in parent group-v284806. [ 956.845608] env[61906]: DEBUG oslo.service.loopingcall [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.845822] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 956.845994] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a4804c2e-0e50-4e67-85e7-46f1d7b35a00 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.866293] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 956.866293] env[61906]: value = "task-1333451" [ 956.866293] env[61906]: _type = "Task" [ 956.866293] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.877639] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333451, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.985393] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52bb2e3b-f020-447b-9a0a-0559903d54b4, 'name': SearchDatastore_Task, 'duration_secs': 0.00824} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.986209] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fab60087-adc0-4840-9edc-d9cf8dc52b31 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.991731] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 956.991731] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52f1cbdd-0113-3270-7bf9-c35d52e32b61" [ 956.991731] env[61906]: _type = "Task" [ 956.991731] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.999322] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52f1cbdd-0113-3270-7bf9-c35d52e32b61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.047687] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.363s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.047687] env[61906]: DEBUG nova.compute.manager [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 957.050842] env[61906]: DEBUG oslo_concurrency.lockutils [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.631s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.051235] env[61906]: DEBUG nova.objects.instance [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lazy-loading 'resources' on Instance uuid ad2db007-e7aa-4c66-9020-ac4acfcff25a {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.087800] env[61906]: DEBUG nova.network.neutron [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Successfully updated port: b5eb6e7b-4b1c-411f-8d99-ad4888193eda {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 957.129917] env[61906]: DEBUG nova.compute.manager [req-6bb1c719-3f84-43bb-b325-0cc20029479f req-c33ca60d-935e-473a-8368-ae2cae912ea0 service nova] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Received event network-changed-ff7dc3ad-711f-4510-8a80-43f9717aeeba {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 957.130141] env[61906]: DEBUG nova.compute.manager [req-6bb1c719-3f84-43bb-b325-0cc20029479f req-c33ca60d-935e-473a-8368-ae2cae912ea0 service nova] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Refreshing instance network info cache due to event network-changed-ff7dc3ad-711f-4510-8a80-43f9717aeeba. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 957.130378] env[61906]: DEBUG oslo_concurrency.lockutils [req-6bb1c719-3f84-43bb-b325-0cc20029479f req-c33ca60d-935e-473a-8368-ae2cae912ea0 service nova] Acquiring lock "refresh_cache-7142e219-6150-4c6f-9632-cbf489035431" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.130674] env[61906]: DEBUG oslo_concurrency.lockutils [req-6bb1c719-3f84-43bb-b325-0cc20029479f req-c33ca60d-935e-473a-8368-ae2cae912ea0 service nova] Acquired lock "refresh_cache-7142e219-6150-4c6f-9632-cbf489035431" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.130935] env[61906]: DEBUG nova.network.neutron [req-6bb1c719-3f84-43bb-b325-0cc20029479f req-c33ca60d-935e-473a-8368-ae2cae912ea0 service nova] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Refreshing network info cache for port ff7dc3ad-711f-4510-8a80-43f9717aeeba {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 957.379090] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333451, 'name': CreateVM_Task} progress is 25%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.501774] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52f1cbdd-0113-3270-7bf9-c35d52e32b61, 'name': SearchDatastore_Task, 'duration_secs': 0.015687} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.502061] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.502328] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 150bf47d-7da9-4e13-ad00-7a8a25b9504e/150bf47d-7da9-4e13-ad00-7a8a25b9504e.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 957.502616] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cd7fa262-0a4c-47cd-85b4-6400434ac145 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.509271] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 957.509271] env[61906]: value = "task-1333452" [ 957.509271] env[61906]: _type = "Task" [ 957.509271] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.517046] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333452, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.533068] env[61906]: DEBUG nova.objects.instance [None req-604fb4d7-9d64-450d-81e8-6cab5a012c75 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lazy-loading 'flavor' on Instance uuid b1159533-c970-49d6-af42-b954b20d92fb {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.554450] env[61906]: DEBUG nova.compute.utils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 957.558633] env[61906]: DEBUG nova.compute.manager [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 957.558809] env[61906]: DEBUG nova.network.neutron [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 957.590529] env[61906]: DEBUG oslo_concurrency.lockutils [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "refresh_cache-03d3fb1b-0e46-4544-b01d-498a2baf3b45" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.591047] env[61906]: DEBUG oslo_concurrency.lockutils [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquired lock "refresh_cache-03d3fb1b-0e46-4544-b01d-498a2baf3b45" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.591047] env[61906]: DEBUG nova.network.neutron [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 957.600107] env[61906]: DEBUG nova.policy [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '23ff59c262484422a8badcbca5a103f9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '76ca0b5a4fd24ede8eac4851fc365848', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 957.785195] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d468dba-0dde-451c-8c46-a978c4e39148 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.795075] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d472897-74c7-4438-99d1-8fda39cf27c6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.832104] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb20c60d-757c-470b-87ba-9e3416dba915 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.841248] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a730aa6-d581-4dcd-9413-1a24af1d35da {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.857031] env[61906]: DEBUG nova.compute.provider_tree [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 957.878455] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333451, 'name': CreateVM_Task} progress is 25%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.903265] env[61906]: DEBUG nova.network.neutron [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Successfully created port: c45bb207-8a87-4c79-8247-704fd5fd810b {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 957.965196] env[61906]: DEBUG nova.network.neutron [req-6bb1c719-3f84-43bb-b325-0cc20029479f req-c33ca60d-935e-473a-8368-ae2cae912ea0 service nova] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Updated VIF entry in instance network info cache for port ff7dc3ad-711f-4510-8a80-43f9717aeeba. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 957.965609] env[61906]: DEBUG nova.network.neutron [req-6bb1c719-3f84-43bb-b325-0cc20029479f req-c33ca60d-935e-473a-8368-ae2cae912ea0 service nova] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Updating instance_info_cache with network_info: [{"id": "ff7dc3ad-711f-4510-8a80-43f9717aeeba", "address": "fa:16:3e:ef:52:74", "network": {"id": "0a3274f7-785f-45e2-9582-0b611955b024", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-593827400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76ca0b5a4fd24ede8eac4851fc365848", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff7dc3ad-71", "ovs_interfaceid": "ff7dc3ad-711f-4510-8a80-43f9717aeeba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.019374] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333452, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.038824] env[61906]: DEBUG oslo_concurrency.lockutils [None req-604fb4d7-9d64-450d-81e8-6cab5a012c75 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "refresh_cache-b1159533-c970-49d6-af42-b954b20d92fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.039052] env[61906]: DEBUG oslo_concurrency.lockutils [None req-604fb4d7-9d64-450d-81e8-6cab5a012c75 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquired lock "refresh_cache-b1159533-c970-49d6-af42-b954b20d92fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.039210] env[61906]: DEBUG nova.network.neutron [None req-604fb4d7-9d64-450d-81e8-6cab5a012c75 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 958.039395] env[61906]: DEBUG nova.objects.instance [None req-604fb4d7-9d64-450d-81e8-6cab5a012c75 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lazy-loading 'info_cache' on Instance uuid b1159533-c970-49d6-af42-b954b20d92fb {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 958.059093] env[61906]: DEBUG nova.compute.manager [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 958.123420] env[61906]: DEBUG nova.network.neutron [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 958.294662] env[61906]: DEBUG nova.network.neutron [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Updating instance_info_cache with network_info: [{"id": "b5eb6e7b-4b1c-411f-8d99-ad4888193eda", "address": "fa:16:3e:fb:4f:d8", "network": {"id": "d123fdf4-6bb4-400b-a0bb-9d477672af2c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2099881656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d2198383d74dac90aa9727a62176df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5eb6e7b-4b", "ovs_interfaceid": "b5eb6e7b-4b1c-411f-8d99-ad4888193eda", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.362051] env[61906]: DEBUG nova.scheduler.client.report [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 958.377713] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333451, 'name': CreateVM_Task, 'duration_secs': 1.346724} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.378430] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 958.379126] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.379307] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.379641] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 958.380108] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6b147c1-bdf6-4d55-bac1-05fcfe0d0ed6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.384698] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 958.384698] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]529e0978-52a0-52e0-f949-ee528cc604ba" [ 958.384698] env[61906]: _type = "Task" [ 958.384698] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.392846] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]529e0978-52a0-52e0-f949-ee528cc604ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.469139] env[61906]: DEBUG oslo_concurrency.lockutils [req-6bb1c719-3f84-43bb-b325-0cc20029479f req-c33ca60d-935e-473a-8368-ae2cae912ea0 service nova] Releasing lock "refresh_cache-7142e219-6150-4c6f-9632-cbf489035431" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.469428] env[61906]: DEBUG nova.compute.manager [req-6bb1c719-3f84-43bb-b325-0cc20029479f req-c33ca60d-935e-473a-8368-ae2cae912ea0 service nova] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Received event network-vif-plugged-b5eb6e7b-4b1c-411f-8d99-ad4888193eda {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 958.469632] env[61906]: DEBUG oslo_concurrency.lockutils [req-6bb1c719-3f84-43bb-b325-0cc20029479f req-c33ca60d-935e-473a-8368-ae2cae912ea0 service nova] Acquiring lock "03d3fb1b-0e46-4544-b01d-498a2baf3b45-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.469841] env[61906]: DEBUG oslo_concurrency.lockutils [req-6bb1c719-3f84-43bb-b325-0cc20029479f req-c33ca60d-935e-473a-8368-ae2cae912ea0 service nova] Lock "03d3fb1b-0e46-4544-b01d-498a2baf3b45-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.470010] env[61906]: DEBUG oslo_concurrency.lockutils [req-6bb1c719-3f84-43bb-b325-0cc20029479f req-c33ca60d-935e-473a-8368-ae2cae912ea0 service nova] Lock "03d3fb1b-0e46-4544-b01d-498a2baf3b45-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.470227] env[61906]: DEBUG nova.compute.manager [req-6bb1c719-3f84-43bb-b325-0cc20029479f req-c33ca60d-935e-473a-8368-ae2cae912ea0 service nova] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] No waiting events found dispatching network-vif-plugged-b5eb6e7b-4b1c-411f-8d99-ad4888193eda {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 958.471054] env[61906]: WARNING nova.compute.manager [req-6bb1c719-3f84-43bb-b325-0cc20029479f req-c33ca60d-935e-473a-8368-ae2cae912ea0 service nova] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Received unexpected event network-vif-plugged-b5eb6e7b-4b1c-411f-8d99-ad4888193eda for instance with vm_state building and task_state spawning. [ 958.519729] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333452, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.57542} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.520009] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 150bf47d-7da9-4e13-ad00-7a8a25b9504e/150bf47d-7da9-4e13-ad00-7a8a25b9504e.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 958.520250] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 958.520597] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cf9ce4b1-9fa1-4dee-80c5-b2261ac41618 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.527401] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 958.527401] env[61906]: value = "task-1333453" [ 958.527401] env[61906]: _type = "Task" [ 958.527401] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.535148] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333453, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.542554] env[61906]: DEBUG nova.objects.base [None req-604fb4d7-9d64-450d-81e8-6cab5a012c75 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 958.798054] env[61906]: DEBUG oslo_concurrency.lockutils [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Releasing lock "refresh_cache-03d3fb1b-0e46-4544-b01d-498a2baf3b45" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.798420] env[61906]: DEBUG nova.compute.manager [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Instance network_info: |[{"id": "b5eb6e7b-4b1c-411f-8d99-ad4888193eda", "address": "fa:16:3e:fb:4f:d8", "network": {"id": "d123fdf4-6bb4-400b-a0bb-9d477672af2c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2099881656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d2198383d74dac90aa9727a62176df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5eb6e7b-4b", "ovs_interfaceid": "b5eb6e7b-4b1c-411f-8d99-ad4888193eda", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 958.798808] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fb:4f:d8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cba18f15-a919-422e-a423-1e705e233389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b5eb6e7b-4b1c-411f-8d99-ad4888193eda', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 958.806211] env[61906]: DEBUG oslo.service.loopingcall [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 958.806448] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 958.806680] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-68df80fa-da74-4935-8539-16bb39259f45 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.826266] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 958.826266] env[61906]: value = "task-1333454" [ 958.826266] env[61906]: _type = "Task" [ 958.826266] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.837274] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333454, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.866912] env[61906]: DEBUG oslo_concurrency.lockutils [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.816s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.869453] env[61906]: DEBUG oslo_concurrency.lockutils [None req-519db265-c6fa-45ee-bd2b-60fdb5fe0763 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 7.343s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.895613] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]529e0978-52a0-52e0-f949-ee528cc604ba, 'name': SearchDatastore_Task, 'duration_secs': 0.013434} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.896631] env[61906]: INFO nova.scheduler.client.report [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Deleted allocations for instance ad2db007-e7aa-4c66-9020-ac4acfcff25a [ 958.897625] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.897867] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 958.898120] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.898276] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.898459] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 958.901109] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b96fa9f5-51d2-4f07-8683-0a9f6709373a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.909835] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 958.910030] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 958.910772] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72b1a9ae-59c6-4387-95bb-cfd884a10312 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.916008] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 958.916008] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52401308-a16a-9831-c19a-87da4c38ef43" [ 958.916008] env[61906]: _type = "Task" [ 958.916008] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.924313] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52401308-a16a-9831-c19a-87da4c38ef43, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.037967] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333453, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071313} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.038261] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 959.039077] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5328cc6d-8154-47f5-8e60-d4e54aa7bcba {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.064265] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] 150bf47d-7da9-4e13-ad00-7a8a25b9504e/150bf47d-7da9-4e13-ad00-7a8a25b9504e.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 959.064569] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9251d0e-e36d-42c3-8e51-ce571551c176 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.080641] env[61906]: DEBUG nova.compute.manager [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 959.089206] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 959.089206] env[61906]: value = "task-1333455" [ 959.089206] env[61906]: _type = "Task" [ 959.089206] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.097284] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333455, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.105217] env[61906]: DEBUG nova.virt.hardware [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 959.105462] env[61906]: DEBUG nova.virt.hardware [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 959.105623] env[61906]: DEBUG nova.virt.hardware [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 959.105813] env[61906]: DEBUG nova.virt.hardware [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 959.106033] env[61906]: DEBUG nova.virt.hardware [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 959.106120] env[61906]: DEBUG nova.virt.hardware [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 959.106335] env[61906]: DEBUG nova.virt.hardware [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 959.106496] env[61906]: DEBUG nova.virt.hardware [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 959.106666] env[61906]: DEBUG nova.virt.hardware [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 959.106834] env[61906]: DEBUG nova.virt.hardware [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 959.107016] env[61906]: DEBUG nova.virt.hardware [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 959.107817] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d086ad77-d1cb-4a37-9046-569190649a24 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.115940] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8458d831-88aa-452a-b4a6-3ecaf65463ee {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.164457] env[61906]: DEBUG nova.compute.manager [req-0df603b7-7247-43d9-9990-2f48ad0cc3de req-bd18a93d-a548-4b72-b94a-4423ef36c3b1 service nova] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Received event network-changed-b5eb6e7b-4b1c-411f-8d99-ad4888193eda {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 959.164676] env[61906]: DEBUG nova.compute.manager [req-0df603b7-7247-43d9-9990-2f48ad0cc3de req-bd18a93d-a548-4b72-b94a-4423ef36c3b1 service nova] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Refreshing instance network info cache due to event network-changed-b5eb6e7b-4b1c-411f-8d99-ad4888193eda. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 959.165149] env[61906]: DEBUG oslo_concurrency.lockutils [req-0df603b7-7247-43d9-9990-2f48ad0cc3de req-bd18a93d-a548-4b72-b94a-4423ef36c3b1 service nova] Acquiring lock "refresh_cache-03d3fb1b-0e46-4544-b01d-498a2baf3b45" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.165149] env[61906]: DEBUG oslo_concurrency.lockutils [req-0df603b7-7247-43d9-9990-2f48ad0cc3de req-bd18a93d-a548-4b72-b94a-4423ef36c3b1 service nova] Acquired lock "refresh_cache-03d3fb1b-0e46-4544-b01d-498a2baf3b45" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.165331] env[61906]: DEBUG nova.network.neutron [req-0df603b7-7247-43d9-9990-2f48ad0cc3de req-bd18a93d-a548-4b72-b94a-4423ef36c3b1 service nova] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Refreshing network info cache for port b5eb6e7b-4b1c-411f-8d99-ad4888193eda {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 959.325216] env[61906]: DEBUG nova.network.neutron [None req-604fb4d7-9d64-450d-81e8-6cab5a012c75 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Updating instance_info_cache with network_info: [{"id": "ba5a12df-7917-4b2d-ac9e-7ef1380dabf7", "address": "fa:16:3e:d3:84:8a", "network": {"id": "1c55d0cc-4c30-4b89-8736-3e96c5c9c507", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-245940583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fbdcd2dc0d04e83b9f82a19bdfd0023", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bcd9d2d-25c8-41ad-9a4a-93b9029ba993", "external-id": "nsx-vlan-transportzone-734", "segmentation_id": 734, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba5a12df-79", "ovs_interfaceid": "ba5a12df-7917-4b2d-ac9e-7ef1380dabf7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.336140] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333454, 'name': CreateVM_Task, 'duration_secs': 0.329472} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.336964] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 959.337702] env[61906]: DEBUG oslo_concurrency.lockutils [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.337882] env[61906]: DEBUG oslo_concurrency.lockutils [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.338219] env[61906]: DEBUG oslo_concurrency.lockutils [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 959.338716] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcdd7b98-9943-4d15-9d98-d76747ee3913 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.343319] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 959.343319] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52a696c3-6a7e-eb6f-8d79-8db5f6f674aa" [ 959.343319] env[61906]: _type = "Task" [ 959.343319] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.351653] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52a696c3-6a7e-eb6f-8d79-8db5f6f674aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.409471] env[61906]: DEBUG oslo_concurrency.lockutils [None req-15e3770a-bea5-4db8-b2f8-da167214d3a5 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "ad2db007-e7aa-4c66-9020-ac4acfcff25a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.422s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.429125] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52401308-a16a-9831-c19a-87da4c38ef43, 'name': SearchDatastore_Task, 'duration_secs': 0.00931} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.429921] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0725b13-b601-441e-8949-86b47dd99afc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.435012] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 959.435012] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52130cb9-c70e-d8f2-eadb-7f95ff50e31d" [ 959.435012] env[61906]: _type = "Task" [ 959.435012] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.445922] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52130cb9-c70e-d8f2-eadb-7f95ff50e31d, 'name': SearchDatastore_Task, 'duration_secs': 0.008775} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.447776] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.447776] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 7142e219-6150-4c6f-9632-cbf489035431/7142e219-6150-4c6f-9632-cbf489035431.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 959.447776] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-88ab531a-e2cc-4091-8ac2-8d40d8373f14 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.453279] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 959.453279] env[61906]: value = "task-1333456" [ 959.453279] env[61906]: _type = "Task" [ 959.453279] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.460733] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333456, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.577759] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d57ed3dd-3ff7-4e1c-a1c7-8c18788dc36d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.585263] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a935f3-77b3-437d-b5cd-0279eabbace2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.619485] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-576f07b0-35f0-430c-83ac-c27b2479ef08 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.626016] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333455, 'name': ReconfigVM_Task, 'duration_secs': 0.292432} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.628174] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Reconfigured VM instance instance-00000053 to attach disk [datastore1] 150bf47d-7da9-4e13-ad00-7a8a25b9504e/150bf47d-7da9-4e13-ad00-7a8a25b9504e.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 959.628863] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4a3d188f-15c8-44b1-8844-80abdefcc7f9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.631426] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80310440-2fdd-4eb0-b34c-33ef7999bf5e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.639618] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 959.639618] env[61906]: value = "task-1333457" [ 959.639618] env[61906]: _type = "Task" [ 959.639618] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.650396] env[61906]: DEBUG nova.compute.provider_tree [None req-519db265-c6fa-45ee-bd2b-60fdb5fe0763 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.657285] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333457, 'name': Rename_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.719198] env[61906]: DEBUG nova.network.neutron [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Successfully updated port: c45bb207-8a87-4c79-8247-704fd5fd810b {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 959.828219] env[61906]: DEBUG oslo_concurrency.lockutils [None req-604fb4d7-9d64-450d-81e8-6cab5a012c75 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Releasing lock "refresh_cache-b1159533-c970-49d6-af42-b954b20d92fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.853946] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52a696c3-6a7e-eb6f-8d79-8db5f6f674aa, 'name': SearchDatastore_Task, 'duration_secs': 0.007937} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.854299] env[61906]: DEBUG oslo_concurrency.lockutils [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.854857] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 959.854857] env[61906]: DEBUG oslo_concurrency.lockutils [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.854996] env[61906]: DEBUG oslo_concurrency.lockutils [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.855096] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 959.855370] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ef0cff2d-8e58-4d8d-bf09-abda063c177e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.868264] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 959.868499] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 959.869308] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b442f3f-940a-4fc1-83e9-bc8442d59132 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.875461] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 959.875461] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52df195a-9c83-0dfa-9798-f1bab21a8a77" [ 959.875461] env[61906]: _type = "Task" [ 959.875461] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.885922] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52df195a-9c83-0dfa-9798-f1bab21a8a77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.919095] env[61906]: DEBUG nova.network.neutron [req-0df603b7-7247-43d9-9990-2f48ad0cc3de req-bd18a93d-a548-4b72-b94a-4423ef36c3b1 service nova] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Updated VIF entry in instance network info cache for port b5eb6e7b-4b1c-411f-8d99-ad4888193eda. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 959.919771] env[61906]: DEBUG nova.network.neutron [req-0df603b7-7247-43d9-9990-2f48ad0cc3de req-bd18a93d-a548-4b72-b94a-4423ef36c3b1 service nova] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Updating instance_info_cache with network_info: [{"id": "b5eb6e7b-4b1c-411f-8d99-ad4888193eda", "address": "fa:16:3e:fb:4f:d8", "network": {"id": "d123fdf4-6bb4-400b-a0bb-9d477672af2c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2099881656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d2198383d74dac90aa9727a62176df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5eb6e7b-4b", "ovs_interfaceid": "b5eb6e7b-4b1c-411f-8d99-ad4888193eda", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.964378] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333456, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.150734] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333457, 'name': Rename_Task, 'duration_secs': 0.431321} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.151038] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 960.151267] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7fca8f93-70f4-4f14-a8b6-f3a5a6bfec2f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.154072] env[61906]: DEBUG nova.scheduler.client.report [None req-519db265-c6fa-45ee-bd2b-60fdb5fe0763 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 960.161785] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 960.161785] env[61906]: value = "task-1333458" [ 960.161785] env[61906]: _type = "Task" [ 960.161785] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.169087] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333458, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.222148] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "refresh_cache-bddfdc7f-09d7-4887-b330-34a596ffa562" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.222273] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquired lock "refresh_cache-bddfdc7f-09d7-4887-b330-34a596ffa562" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.222397] env[61906]: DEBUG nova.network.neutron [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 960.332850] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-604fb4d7-9d64-450d-81e8-6cab5a012c75 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 960.332850] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d6934d76-9d4c-4191-8351-d2f51130cc9a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.340377] env[61906]: DEBUG oslo_vmware.api [None req-604fb4d7-9d64-450d-81e8-6cab5a012c75 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 960.340377] env[61906]: value = "task-1333459" [ 960.340377] env[61906]: _type = "Task" [ 960.340377] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.347972] env[61906]: DEBUG oslo_vmware.api [None req-604fb4d7-9d64-450d-81e8-6cab5a012c75 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333459, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.385210] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52df195a-9c83-0dfa-9798-f1bab21a8a77, 'name': SearchDatastore_Task, 'duration_secs': 0.070815} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.386031] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-821af933-addd-4dbd-831b-efa847e06a4b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.391365] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 960.391365] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52ee3dcb-fa2a-83f3-28d1-7d468c2cc029" [ 960.391365] env[61906]: _type = "Task" [ 960.391365] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.399182] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52ee3dcb-fa2a-83f3-28d1-7d468c2cc029, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.423035] env[61906]: DEBUG oslo_concurrency.lockutils [req-0df603b7-7247-43d9-9990-2f48ad0cc3de req-bd18a93d-a548-4b72-b94a-4423ef36c3b1 service nova] Releasing lock "refresh_cache-03d3fb1b-0e46-4544-b01d-498a2baf3b45" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.463998] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333456, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.534957} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.464277] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 7142e219-6150-4c6f-9632-cbf489035431/7142e219-6150-4c6f-9632-cbf489035431.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 960.464496] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 960.464749] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1cdc26d0-f1ca-415c-8a38-c20dea6fd74d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.471447] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 960.471447] env[61906]: value = "task-1333460" [ 960.471447] env[61906]: _type = "Task" [ 960.471447] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.479140] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333460, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.671419] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333458, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.761075] env[61906]: DEBUG nova.network.neutron [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 960.851724] env[61906]: DEBUG oslo_vmware.api [None req-604fb4d7-9d64-450d-81e8-6cab5a012c75 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333459, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.901851] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52ee3dcb-fa2a-83f3-28d1-7d468c2cc029, 'name': SearchDatastore_Task, 'duration_secs': 0.009706} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.904209] env[61906]: DEBUG oslo_concurrency.lockutils [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.904556] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 03d3fb1b-0e46-4544-b01d-498a2baf3b45/03d3fb1b-0e46-4544-b01d-498a2baf3b45.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 960.904759] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7465bace-3964-4d38-8c56-094b58bf295b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.911575] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 960.911575] env[61906]: value = "task-1333461" [ 960.911575] env[61906]: _type = "Task" [ 960.911575] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.918582] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333461, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.939207] env[61906]: DEBUG nova.network.neutron [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Updating instance_info_cache with network_info: [{"id": "c45bb207-8a87-4c79-8247-704fd5fd810b", "address": "fa:16:3e:90:fd:33", "network": {"id": "0a3274f7-785f-45e2-9582-0b611955b024", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-593827400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76ca0b5a4fd24ede8eac4851fc365848", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc45bb207-8a", "ovs_interfaceid": "c45bb207-8a87-4c79-8247-704fd5fd810b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.980396] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333460, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.167155} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.980685] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 960.981441] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5eb7c70-d612-4bad-a726-f7eb88d29f42 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.003876] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] 7142e219-6150-4c6f-9632-cbf489035431/7142e219-6150-4c6f-9632-cbf489035431.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 961.004126] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e074c7d3-70dd-4885-bd42-62bb687fd2de {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.037912] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 961.037912] env[61906]: value = "task-1333462" [ 961.037912] env[61906]: _type = "Task" [ 961.037912] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.046491] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333462, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.163349] env[61906]: DEBUG oslo_concurrency.lockutils [None req-519db265-c6fa-45ee-bd2b-60fdb5fe0763 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.294s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.167285] env[61906]: DEBUG oslo_concurrency.lockutils [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.604s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.167599] env[61906]: DEBUG nova.objects.instance [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Lazy-loading 'resources' on Instance uuid 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 961.179288] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333458, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.197552] env[61906]: DEBUG nova.compute.manager [req-6906b0c1-8264-462e-af99-ac2bdbc7218d req-3712967d-6d1c-42a0-9bbf-db7f00b090e1 service nova] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Received event network-vif-plugged-c45bb207-8a87-4c79-8247-704fd5fd810b {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 961.197931] env[61906]: DEBUG oslo_concurrency.lockutils [req-6906b0c1-8264-462e-af99-ac2bdbc7218d req-3712967d-6d1c-42a0-9bbf-db7f00b090e1 service nova] Acquiring lock "bddfdc7f-09d7-4887-b330-34a596ffa562-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.197931] env[61906]: DEBUG oslo_concurrency.lockutils [req-6906b0c1-8264-462e-af99-ac2bdbc7218d req-3712967d-6d1c-42a0-9bbf-db7f00b090e1 service nova] Lock "bddfdc7f-09d7-4887-b330-34a596ffa562-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.198362] env[61906]: DEBUG oslo_concurrency.lockutils [req-6906b0c1-8264-462e-af99-ac2bdbc7218d req-3712967d-6d1c-42a0-9bbf-db7f00b090e1 service nova] Lock "bddfdc7f-09d7-4887-b330-34a596ffa562-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.198362] env[61906]: DEBUG nova.compute.manager [req-6906b0c1-8264-462e-af99-ac2bdbc7218d req-3712967d-6d1c-42a0-9bbf-db7f00b090e1 service nova] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] No waiting events found dispatching network-vif-plugged-c45bb207-8a87-4c79-8247-704fd5fd810b {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 961.198757] env[61906]: WARNING nova.compute.manager [req-6906b0c1-8264-462e-af99-ac2bdbc7218d req-3712967d-6d1c-42a0-9bbf-db7f00b090e1 service nova] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Received unexpected event network-vif-plugged-c45bb207-8a87-4c79-8247-704fd5fd810b for instance with vm_state building and task_state spawning. [ 961.198757] env[61906]: DEBUG nova.compute.manager [req-6906b0c1-8264-462e-af99-ac2bdbc7218d req-3712967d-6d1c-42a0-9bbf-db7f00b090e1 service nova] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Received event network-changed-c45bb207-8a87-4c79-8247-704fd5fd810b {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 961.199062] env[61906]: DEBUG nova.compute.manager [req-6906b0c1-8264-462e-af99-ac2bdbc7218d req-3712967d-6d1c-42a0-9bbf-db7f00b090e1 service nova] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Refreshing instance network info cache due to event network-changed-c45bb207-8a87-4c79-8247-704fd5fd810b. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 961.199125] env[61906]: DEBUG oslo_concurrency.lockutils [req-6906b0c1-8264-462e-af99-ac2bdbc7218d req-3712967d-6d1c-42a0-9bbf-db7f00b090e1 service nova] Acquiring lock "refresh_cache-bddfdc7f-09d7-4887-b330-34a596ffa562" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.353114] env[61906]: DEBUG oslo_vmware.api [None req-604fb4d7-9d64-450d-81e8-6cab5a012c75 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333459, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.422202] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333461, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.442186] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Releasing lock "refresh_cache-bddfdc7f-09d7-4887-b330-34a596ffa562" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.442576] env[61906]: DEBUG nova.compute.manager [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Instance network_info: |[{"id": "c45bb207-8a87-4c79-8247-704fd5fd810b", "address": "fa:16:3e:90:fd:33", "network": {"id": "0a3274f7-785f-45e2-9582-0b611955b024", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-593827400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76ca0b5a4fd24ede8eac4851fc365848", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc45bb207-8a", "ovs_interfaceid": "c45bb207-8a87-4c79-8247-704fd5fd810b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 961.442898] env[61906]: DEBUG oslo_concurrency.lockutils [req-6906b0c1-8264-462e-af99-ac2bdbc7218d req-3712967d-6d1c-42a0-9bbf-db7f00b090e1 service nova] Acquired lock "refresh_cache-bddfdc7f-09d7-4887-b330-34a596ffa562" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.443103] env[61906]: DEBUG nova.network.neutron [req-6906b0c1-8264-462e-af99-ac2bdbc7218d req-3712967d-6d1c-42a0-9bbf-db7f00b090e1 service nova] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Refreshing network info cache for port c45bb207-8a87-4c79-8247-704fd5fd810b {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 961.444497] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:90:fd:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c45bb207-8a87-4c79-8247-704fd5fd810b', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 961.452220] env[61906]: DEBUG oslo.service.loopingcall [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.453332] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 961.453570] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eabd4df1-2ab7-41d0-bb1e-8a144aa01624 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.475914] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 961.475914] env[61906]: value = "task-1333463" [ 961.475914] env[61906]: _type = "Task" [ 961.475914] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.486071] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333463, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.550602] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333462, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.683167] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333458, 'name': PowerOnVM_Task} progress is 71%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.734241] env[61906]: INFO nova.scheduler.client.report [None req-519db265-c6fa-45ee-bd2b-60fdb5fe0763 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Deleted allocation for migration f037f355-83fb-4668-9837-81e957b39c1e [ 961.850694] env[61906]: DEBUG oslo_vmware.api [None req-604fb4d7-9d64-450d-81e8-6cab5a012c75 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333459, 'name': PowerOnVM_Task, 'duration_secs': 1.179885} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.853114] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-604fb4d7-9d64-450d-81e8-6cab5a012c75 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 961.853370] env[61906]: DEBUG nova.compute.manager [None req-604fb4d7-9d64-450d-81e8-6cab5a012c75 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 961.854592] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dfc87a9-48af-4a92-bed7-cbf9b2d85ac3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.863657] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6be21341-8e5e-446f-b1a8-4f5aab3d61f5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.872111] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4de6932c-38e4-446f-980d-5c35349e798f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.907050] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c80f286c-ac49-4c63-a8ce-b3105d97d7fe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.917017] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d677fc59-88a8-407a-92cf-54d6ed9e6e17 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.925717] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333461, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.662677} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.933316] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 03d3fb1b-0e46-4544-b01d-498a2baf3b45/03d3fb1b-0e46-4544-b01d-498a2baf3b45.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 961.933572] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 961.934313] env[61906]: DEBUG nova.compute.provider_tree [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 961.938598] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d6d4dd0b-79f8-4fc1-82f4-a820a6332b7c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.942304] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 961.942304] env[61906]: value = "task-1333464" [ 961.942304] env[61906]: _type = "Task" [ 961.942304] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.950312] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333464, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.988423] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333463, 'name': CreateVM_Task, 'duration_secs': 0.332928} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.990774] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 961.991474] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.991668] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.992049] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 961.993323] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84fa7103-b8eb-4c06-9dbc-414485f65e6c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.997071] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 961.997071] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5259ef4c-1020-2e9c-f4a5-704fbbac0167" [ 961.997071] env[61906]: _type = "Task" [ 961.997071] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.005172] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5259ef4c-1020-2e9c-f4a5-704fbbac0167, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.052220] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333462, 'name': ReconfigVM_Task, 'duration_secs': 0.659188} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.052548] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Reconfigured VM instance instance-00000054 to attach disk [datastore1] 7142e219-6150-4c6f-9632-cbf489035431/7142e219-6150-4c6f-9632-cbf489035431.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 962.053204] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1987c712-c273-4c70-9066-5da3782ee74c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.059710] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 962.059710] env[61906]: value = "task-1333465" [ 962.059710] env[61906]: _type = "Task" [ 962.059710] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.068869] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333465, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.160143] env[61906]: DEBUG nova.network.neutron [req-6906b0c1-8264-462e-af99-ac2bdbc7218d req-3712967d-6d1c-42a0-9bbf-db7f00b090e1 service nova] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Updated VIF entry in instance network info cache for port c45bb207-8a87-4c79-8247-704fd5fd810b. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 962.160668] env[61906]: DEBUG nova.network.neutron [req-6906b0c1-8264-462e-af99-ac2bdbc7218d req-3712967d-6d1c-42a0-9bbf-db7f00b090e1 service nova] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Updating instance_info_cache with network_info: [{"id": "c45bb207-8a87-4c79-8247-704fd5fd810b", "address": "fa:16:3e:90:fd:33", "network": {"id": "0a3274f7-785f-45e2-9582-0b611955b024", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-593827400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76ca0b5a4fd24ede8eac4851fc365848", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc45bb207-8a", "ovs_interfaceid": "c45bb207-8a87-4c79-8247-704fd5fd810b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.178503] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333458, 'name': PowerOnVM_Task} progress is 81%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.239958] env[61906]: DEBUG oslo_concurrency.lockutils [None req-519db265-c6fa-45ee-bd2b-60fdb5fe0763 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 13.607s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.453486] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333464, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064774} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.453792] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 962.454583] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda9d7d4-85a9-43e5-9abf-31f3d1049f90 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.476292] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] 03d3fb1b-0e46-4544-b01d-498a2baf3b45/03d3fb1b-0e46-4544-b01d-498a2baf3b45.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 962.477459] env[61906]: ERROR nova.scheduler.client.report [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [req-7dc6680b-61fd-409e-8c73-534e4638f0eb] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID a5b57df7-a16e-44f9-9b2d-23c518860263. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-7dc6680b-61fd-409e-8c73-534e4638f0eb"}]} [ 962.477805] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-253e7915-a069-4af1-9600-6d8615bf5139 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.502239] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 962.502239] env[61906]: value = "task-1333466" [ 962.502239] env[61906]: _type = "Task" [ 962.502239] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.508662] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5259ef4c-1020-2e9c-f4a5-704fbbac0167, 'name': SearchDatastore_Task, 'duration_secs': 0.008545} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.509311] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.509694] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 962.509813] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.509962] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.510186] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 962.514605] env[61906]: DEBUG nova.scheduler.client.report [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Refreshing inventories for resource provider a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 962.517191] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1b6664fb-e36d-4f4e-9cdb-983d86d8bb97 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.519871] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333466, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.526777] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 962.526941] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 962.527720] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-904bc7f8-2393-4c01-9d43-3cc050f0812c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.534209] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 962.534209] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]525fe003-faf2-e3d8-3995-dab0d26d524f" [ 962.534209] env[61906]: _type = "Task" [ 962.534209] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.537727] env[61906]: DEBUG nova.scheduler.client.report [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Updating ProviderTree inventory for provider a5b57df7-a16e-44f9-9b2d-23c518860263 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 962.537962] env[61906]: DEBUG nova.compute.provider_tree [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 962.544559] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]525fe003-faf2-e3d8-3995-dab0d26d524f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.553272] env[61906]: DEBUG nova.scheduler.client.report [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Refreshing aggregate associations for resource provider a5b57df7-a16e-44f9-9b2d-23c518860263, aggregates: None {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 962.568657] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333465, 'name': Rename_Task, 'duration_secs': 0.196134} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.569143] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 962.569224] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7addabcf-d777-4741-8444-f0ed7fed0173 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.574744] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 962.574744] env[61906]: value = "task-1333467" [ 962.574744] env[61906]: _type = "Task" [ 962.574744] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.575459] env[61906]: DEBUG nova.scheduler.client.report [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Refreshing trait associations for resource provider a5b57df7-a16e-44f9-9b2d-23c518860263, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64 {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 962.585421] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333467, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.664310] env[61906]: DEBUG oslo_concurrency.lockutils [req-6906b0c1-8264-462e-af99-ac2bdbc7218d req-3712967d-6d1c-42a0-9bbf-db7f00b090e1 service nova] Releasing lock "refresh_cache-bddfdc7f-09d7-4887-b330-34a596ffa562" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.677590] env[61906]: DEBUG oslo_vmware.api [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333458, 'name': PowerOnVM_Task, 'duration_secs': 2.07277} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.680245] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 962.680469] env[61906]: INFO nova.compute.manager [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Took 10.71 seconds to spawn the instance on the hypervisor. [ 962.680652] env[61906]: DEBUG nova.compute.manager [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 962.681658] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-216066f1-a1ac-4007-bf09-9389b261458d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.794212] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a9357d7-70a7-4ef9-8264-0006293a2bdb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.801837] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77d4d55c-d7a8-466b-bd41-1df45c4b6b37 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.832108] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92f1fcc3-83ca-40c9-938a-7ba18dbe64ec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.839818] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbf18181-3b1e-44ef-9446-516dd976d840 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.853448] env[61906]: DEBUG nova.compute.provider_tree [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 963.013381] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333466, 'name': ReconfigVM_Task, 'duration_secs': 0.327317} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.013703] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Reconfigured VM instance instance-00000056 to attach disk [datastore1] 03d3fb1b-0e46-4544-b01d-498a2baf3b45/03d3fb1b-0e46-4544-b01d-498a2baf3b45.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 963.014634] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e7d56a0d-22e0-4ffc-a917-a552c75856f3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.037980] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 963.037980] env[61906]: value = "task-1333468" [ 963.037980] env[61906]: _type = "Task" [ 963.037980] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.045048] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]525fe003-faf2-e3d8-3995-dab0d26d524f, 'name': SearchDatastore_Task, 'duration_secs': 0.01345} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.046191] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ddbe87ff-1d51-4c04-9e0a-825235ca7eb3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.052975] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333468, 'name': Rename_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.057337] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 963.057337] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52ea4cbd-e6e8-1f9b-5f50-bd3044a0dbd9" [ 963.057337] env[61906]: _type = "Task" [ 963.057337] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.064699] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52ea4cbd-e6e8-1f9b-5f50-bd3044a0dbd9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.087565] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333467, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.203143] env[61906]: INFO nova.compute.manager [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Took 15.85 seconds to build instance. [ 963.395096] env[61906]: DEBUG nova.scheduler.client.report [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Updated inventory for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with generation 99 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 963.395209] env[61906]: DEBUG nova.compute.provider_tree [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Updating resource provider a5b57df7-a16e-44f9-9b2d-23c518860263 generation from 99 to 100 during operation: update_inventory {{(pid=61906) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 963.395525] env[61906]: DEBUG nova.compute.provider_tree [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 963.439319] env[61906]: DEBUG oslo_concurrency.lockutils [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.439588] env[61906]: DEBUG oslo_concurrency.lockutils [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.439800] env[61906]: DEBUG oslo_concurrency.lockutils [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.439991] env[61906]: DEBUG oslo_concurrency.lockutils [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.440207] env[61906]: DEBUG oslo_concurrency.lockutils [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.443255] env[61906]: INFO nova.compute.manager [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Terminating instance [ 963.446897] env[61906]: DEBUG nova.compute.manager [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 963.447129] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 963.447962] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d820b3c8-b57c-426c-8b3a-bc351fa914e5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.456313] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 963.456648] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6689c3c4-327b-4798-b24b-f7ee0e150555 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.463071] env[61906]: DEBUG oslo_vmware.api [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 963.463071] env[61906]: value = "task-1333469" [ 963.463071] env[61906]: _type = "Task" [ 963.463071] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.471650] env[61906]: DEBUG oslo_vmware.api [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333469, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.552021] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333468, 'name': Rename_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.567857] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52ea4cbd-e6e8-1f9b-5f50-bd3044a0dbd9, 'name': SearchDatastore_Task, 'duration_secs': 0.05139} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.568219] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.568518] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] bddfdc7f-09d7-4887-b330-34a596ffa562/bddfdc7f-09d7-4887-b330-34a596ffa562.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 963.568830] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-07820d2b-faa4-4613-906c-08065bd02157 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.576383] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 963.576383] env[61906]: value = "task-1333470" [ 963.576383] env[61906]: _type = "Task" [ 963.576383] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.587399] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333470, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.590926] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333467, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.658227] env[61906]: DEBUG nova.compute.manager [req-719752c0-8c38-45d5-9154-34114b48e77b req-16b17db9-f851-466f-b137-3c5bd74a83a3 service nova] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Received event network-changed-5c99ae7c-41d1-48de-a53f-fc8d1a6a4ca2 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.658523] env[61906]: DEBUG nova.compute.manager [req-719752c0-8c38-45d5-9154-34114b48e77b req-16b17db9-f851-466f-b137-3c5bd74a83a3 service nova] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Refreshing instance network info cache due to event network-changed-5c99ae7c-41d1-48de-a53f-fc8d1a6a4ca2. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 963.658694] env[61906]: DEBUG oslo_concurrency.lockutils [req-719752c0-8c38-45d5-9154-34114b48e77b req-16b17db9-f851-466f-b137-3c5bd74a83a3 service nova] Acquiring lock "refresh_cache-150bf47d-7da9-4e13-ad00-7a8a25b9504e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.658796] env[61906]: DEBUG oslo_concurrency.lockutils [req-719752c0-8c38-45d5-9154-34114b48e77b req-16b17db9-f851-466f-b137-3c5bd74a83a3 service nova] Acquired lock "refresh_cache-150bf47d-7da9-4e13-ad00-7a8a25b9504e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.658967] env[61906]: DEBUG nova.network.neutron [req-719752c0-8c38-45d5-9154-34114b48e77b req-16b17db9-f851-466f-b137-3c5bd74a83a3 service nova] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Refreshing network info cache for port 5c99ae7c-41d1-48de-a53f-fc8d1a6a4ca2 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 963.703963] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6564f469-eb53-4b3e-a71f-d4d7e4ea0aff tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "150bf47d-7da9-4e13-ad00-7a8a25b9504e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.359s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.900390] env[61906]: DEBUG oslo_concurrency.lockutils [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.733s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.903413] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.937s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.903810] env[61906]: DEBUG nova.objects.instance [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lazy-loading 'resources' on Instance uuid 83b59194-b86f-416a-a7fd-18f75551c2c0 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 963.928500] env[61906]: INFO nova.scheduler.client.report [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Deleted allocations for instance 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5 [ 963.976145] env[61906]: DEBUG oslo_vmware.api [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333469, 'name': PowerOffVM_Task, 'duration_secs': 0.188452} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.976703] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 963.977036] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 963.977549] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7672e3e9-3187-4560-ab16-7c5f272f5cd4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.025658] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "458a1669-a62b-4313-874a-e49809d5c034" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.026075] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "458a1669-a62b-4313-874a-e49809d5c034" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.047792] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 964.048127] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 964.048336] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Deleting the datastore file [datastore1] e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 964.048602] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d56e9838-76b6-4438-8de5-802a0f8e4349 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.054610] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333468, 'name': Rename_Task, 'duration_secs': 0.914594} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.055426] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 964.055754] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e4826b4d-fc72-44a0-a8cc-a39fddb681e6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.059342] env[61906]: DEBUG oslo_vmware.api [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 964.059342] env[61906]: value = "task-1333472" [ 964.059342] env[61906]: _type = "Task" [ 964.059342] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.066969] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 964.066969] env[61906]: value = "task-1333473" [ 964.066969] env[61906]: _type = "Task" [ 964.066969] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.071441] env[61906]: DEBUG oslo_vmware.api [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333472, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.080746] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333473, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.095024] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333470, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.100299] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333467, 'name': PowerOnVM_Task, 'duration_secs': 1.123478} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.100637] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 964.100890] env[61906]: INFO nova.compute.manager [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Took 9.72 seconds to spawn the instance on the hypervisor. [ 964.101338] env[61906]: DEBUG nova.compute.manager [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 964.102131] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb32e81c-e9cf-4f2e-8b22-f2e56fe56a04 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.398071] env[61906]: DEBUG nova.network.neutron [req-719752c0-8c38-45d5-9154-34114b48e77b req-16b17db9-f851-466f-b137-3c5bd74a83a3 service nova] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Updated VIF entry in instance network info cache for port 5c99ae7c-41d1-48de-a53f-fc8d1a6a4ca2. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 964.398426] env[61906]: DEBUG nova.network.neutron [req-719752c0-8c38-45d5-9154-34114b48e77b req-16b17db9-f851-466f-b137-3c5bd74a83a3 service nova] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Updating instance_info_cache with network_info: [{"id": "5c99ae7c-41d1-48de-a53f-fc8d1a6a4ca2", "address": "fa:16:3e:b2:fc:be", "network": {"id": "c3bcef0f-44a7-44f7-8130-df77c0d3ac2d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-330384233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.247", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "129cdb06b47346b6895df45e73f7b2dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35e463c7-7d78-4d66-8efd-6127b1f3ee17", "external-id": "nsx-vlan-transportzone-175", "segmentation_id": 175, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c99ae7c-41", "ovs_interfaceid": "5c99ae7c-41d1-48de-a53f-fc8d1a6a4ca2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.440327] env[61906]: DEBUG oslo_concurrency.lockutils [None req-24e66c35-4ab5-4e48-b6bd-c09483ea9481 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Lock "8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.696s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.528910] env[61906]: DEBUG nova.compute.manager [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 964.571865] env[61906]: DEBUG oslo_vmware.api [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333472, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.580535] env[61906]: DEBUG oslo_vmware.api [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333473, 'name': PowerOnVM_Task, 'duration_secs': 0.474904} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.585510] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 964.585729] env[61906]: INFO nova.compute.manager [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Took 7.88 seconds to spawn the instance on the hypervisor. [ 964.585915] env[61906]: DEBUG nova.compute.manager [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 964.587262] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae40a372-2b29-4d19-9fc3-5dfb2d8e8b0e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.594599] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333470, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.667771} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.596271] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] bddfdc7f-09d7-4887-b330-34a596ffa562/bddfdc7f-09d7-4887-b330-34a596ffa562.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 964.596496] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 964.601419] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-17003474-cd26-449c-8d97-365c6c53fa45 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.608098] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 964.608098] env[61906]: value = "task-1333474" [ 964.608098] env[61906]: _type = "Task" [ 964.608098] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.614174] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05458ac8-ec23-41c3-8d6d-56829eeb3953 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.623151] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333474, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.628897] env[61906]: INFO nova.compute.manager [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Took 14.98 seconds to build instance. [ 964.630563] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fc40e85-a6e1-44a3-ad19-da3d8f57466b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.667044] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27616603-7517-4f3d-b924-a172279ae2b3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.679711] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96b8dd36-5859-4d17-94b4-930432e084e5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.696040] env[61906]: DEBUG nova.compute.provider_tree [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 964.901678] env[61906]: DEBUG oslo_concurrency.lockutils [req-719752c0-8c38-45d5-9154-34114b48e77b req-16b17db9-f851-466f-b137-3c5bd74a83a3 service nova] Releasing lock "refresh_cache-150bf47d-7da9-4e13-ad00-7a8a25b9504e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.047585] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.069262] env[61906]: DEBUG oslo_vmware.api [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333472, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.623062} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.069516] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 965.069706] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 965.069888] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 965.070078] env[61906]: INFO nova.compute.manager [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Took 1.62 seconds to destroy the instance on the hypervisor. [ 965.070327] env[61906]: DEBUG oslo.service.loopingcall [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.070520] env[61906]: DEBUG nova.compute.manager [-] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 965.070634] env[61906]: DEBUG nova.network.neutron [-] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 965.117423] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333474, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.198366} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.117656] env[61906]: INFO nova.compute.manager [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Took 15.46 seconds to build instance. [ 965.118512] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 965.119289] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e70bed25-9461-4115-89fd-8083a4ef6825 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.141944] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] bddfdc7f-09d7-4887-b330-34a596ffa562/bddfdc7f-09d7-4887-b330-34a596ffa562.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 965.146020] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "7142e219-6150-4c6f-9632-cbf489035431" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.511s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.146318] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-36289155-8e91-4fb1-9d09-6dc0c79b89a8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.171692] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 965.171692] env[61906]: value = "task-1333475" [ 965.171692] env[61906]: _type = "Task" [ 965.171692] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.175707] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquiring lock "11cb9d89-2304-45a6-b4bb-fcf8417a0518" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.175992] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Lock "11cb9d89-2304-45a6-b4bb-fcf8417a0518" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.176254] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquiring lock "11cb9d89-2304-45a6-b4bb-fcf8417a0518-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.176447] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Lock "11cb9d89-2304-45a6-b4bb-fcf8417a0518-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.176650] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Lock "11cb9d89-2304-45a6-b4bb-fcf8417a0518-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.178690] env[61906]: INFO nova.compute.manager [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Terminating instance [ 965.181079] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquiring lock "refresh_cache-11cb9d89-2304-45a6-b4bb-fcf8417a0518" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.181246] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquired lock "refresh_cache-11cb9d89-2304-45a6-b4bb-fcf8417a0518" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.181413] env[61906]: DEBUG nova.network.neutron [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 965.185991] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333475, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.199570] env[61906]: DEBUG nova.scheduler.client.report [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 965.622920] env[61906]: DEBUG oslo_concurrency.lockutils [None req-68e883bf-b845-48a1-902d-7e444e49557c tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "03d3fb1b-0e46-4544-b01d-498a2baf3b45" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.973s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.683042] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333475, 'name': ReconfigVM_Task, 'duration_secs': 0.279629} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.683427] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Reconfigured VM instance instance-00000055 to attach disk [datastore1] bddfdc7f-09d7-4887-b330-34a596ffa562/bddfdc7f-09d7-4887-b330-34a596ffa562.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 965.684091] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-49b5f043-e6cb-47e2-86b8-7647820e8fc9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.689635] env[61906]: DEBUG nova.compute.manager [req-04c0b4c9-4751-45b7-966a-4a743034471f req-59221d04-2083-46a5-addc-f3696b2c97c1 service nova] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Received event network-vif-deleted-b54c27f8-7727-4166-af82-5b6fe0d3519e {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.689917] env[61906]: INFO nova.compute.manager [req-04c0b4c9-4751-45b7-966a-4a743034471f req-59221d04-2083-46a5-addc-f3696b2c97c1 service nova] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Neutron deleted interface b54c27f8-7727-4166-af82-5b6fe0d3519e; detaching it from the instance and deleting it from the info cache [ 965.690202] env[61906]: DEBUG nova.network.neutron [req-04c0b4c9-4751-45b7-966a-4a743034471f req-59221d04-2083-46a5-addc-f3696b2c97c1 service nova] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.697410] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 965.697410] env[61906]: value = "task-1333476" [ 965.697410] env[61906]: _type = "Task" [ 965.697410] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.706904] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.804s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.708755] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333476, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.710020] env[61906]: DEBUG nova.network.neutron [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 965.711222] env[61906]: DEBUG oslo_concurrency.lockutils [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.694s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.712265] env[61906]: DEBUG nova.objects.instance [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lazy-loading 'resources' on Instance uuid a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 965.729247] env[61906]: INFO nova.scheduler.client.report [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Deleted allocations for instance 83b59194-b86f-416a-a7fd-18f75551c2c0 [ 965.774237] env[61906]: DEBUG nova.network.neutron [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.850766] env[61906]: DEBUG nova.network.neutron [-] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.172392] env[61906]: DEBUG nova.compute.manager [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 966.173385] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6249d8ae-17a3-4743-872e-3d3e23822557 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.193372] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e2e9cff8-33d7-485d-8db4-e33b5def86c9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.205363] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43c64a9f-4773-4784-ab84-2df2fbb13a69 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.223147] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333476, 'name': Rename_Task, 'duration_secs': 0.150628} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.224529] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 966.224819] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7ec8fb97-302c-45b8-a7c2-d34fbe404e21 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.244639] env[61906]: DEBUG nova.compute.manager [req-04c0b4c9-4751-45b7-966a-4a743034471f req-59221d04-2083-46a5-addc-f3696b2c97c1 service nova] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Detach interface failed, port_id=b54c27f8-7727-4166-af82-5b6fe0d3519e, reason: Instance e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 966.246495] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fe7e4355-84a9-451a-aab7-eac2f499ee0d tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "83b59194-b86f-416a-a7fd-18f75551c2c0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.702s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.249847] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 966.249847] env[61906]: value = "task-1333477" [ 966.249847] env[61906]: _type = "Task" [ 966.249847] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.258381] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333477, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.277333] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Releasing lock "refresh_cache-11cb9d89-2304-45a6-b4bb-fcf8417a0518" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.278137] env[61906]: DEBUG nova.compute.manager [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 966.278137] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 966.281702] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbc51f26-8db3-4ef5-ba69-26c1440809e5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.289953] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 966.290295] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-641ec2d4-9e14-4495-9e84-2df0886b4b79 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.299046] env[61906]: DEBUG oslo_vmware.api [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 966.299046] env[61906]: value = "task-1333478" [ 966.299046] env[61906]: _type = "Task" [ 966.299046] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.306675] env[61906]: DEBUG oslo_vmware.api [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333478, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.354151] env[61906]: INFO nova.compute.manager [-] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Took 1.28 seconds to deallocate network for instance. [ 966.427330] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02abe8e2-ff2c-4b14-94a0-5dbeda9aa7df {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.438410] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-270f1f60-c61d-4dfe-9d6a-c33bcd67d773 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.474912] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-335a7986-24a8-4645-b191-4ed2eb6fe561 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.482601] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78a948b0-4d56-4360-b588-eab169fb1720 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.498690] env[61906]: DEBUG nova.compute.provider_tree [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 966.684435] env[61906]: INFO nova.compute.manager [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] instance snapshotting [ 966.687524] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e29b3ba0-807c-44b3-9e2c-35b57b7d2768 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.707203] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1408ab05-8d9d-43b1-88f9-3dada5c8b7f1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.760563] env[61906]: DEBUG oslo_vmware.api [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333477, 'name': PowerOnVM_Task, 'duration_secs': 0.483881} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.760839] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 966.761058] env[61906]: INFO nova.compute.manager [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Took 7.68 seconds to spawn the instance on the hypervisor. [ 966.761244] env[61906]: DEBUG nova.compute.manager [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 966.762007] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7a9edc7-524f-4e0a-a3dc-ab2505b90814 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.808375] env[61906]: DEBUG oslo_vmware.api [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333478, 'name': PowerOffVM_Task, 'duration_secs': 0.214583} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.808651] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 966.808830] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 966.809099] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fe2c907a-dbae-4319-9f6d-b2a4b40ce3ff {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.832208] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 966.832504] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 966.832658] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Deleting the datastore file [datastore1] 11cb9d89-2304-45a6-b4bb-fcf8417a0518 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 966.832982] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5a2ab386-c62e-468e-97e7-10e2ab3e5e68 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.839906] env[61906]: DEBUG oslo_vmware.api [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for the task: (returnval){ [ 966.839906] env[61906]: value = "task-1333480" [ 966.839906] env[61906]: _type = "Task" [ 966.839906] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.848195] env[61906]: DEBUG oslo_vmware.api [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333480, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.863462] env[61906]: DEBUG oslo_concurrency.lockutils [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.001906] env[61906]: DEBUG nova.scheduler.client.report [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 967.218892] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Creating Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 967.220765] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-708680cc-c1f5-46ea-86e0-475f6250133b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.225343] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "a3e3ad00-6921-4072-8cb1-d80302883513" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.225577] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "a3e3ad00-6921-4072-8cb1-d80302883513" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.232375] env[61906]: DEBUG oslo_vmware.api [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 967.232375] env[61906]: value = "task-1333481" [ 967.232375] env[61906]: _type = "Task" [ 967.232375] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.241666] env[61906]: DEBUG oslo_vmware.api [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333481, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.277644] env[61906]: INFO nova.compute.manager [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Took 17.55 seconds to build instance. [ 967.351425] env[61906]: DEBUG oslo_vmware.api [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Task: {'id': task-1333480, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.23282} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.351700] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 967.351909] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 967.352121] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 967.352350] env[61906]: INFO nova.compute.manager [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Took 1.07 seconds to destroy the instance on the hypervisor. [ 967.352660] env[61906]: DEBUG oslo.service.loopingcall [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 967.352883] env[61906]: DEBUG nova.compute.manager [-] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 967.352992] env[61906]: DEBUG nova.network.neutron [-] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 967.369308] env[61906]: DEBUG nova.network.neutron [-] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 967.507335] env[61906]: DEBUG oslo_concurrency.lockutils [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.796s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.511032] env[61906]: DEBUG oslo_concurrency.lockutils [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.787s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.512273] env[61906]: INFO nova.compute.claims [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 967.526629] env[61906]: INFO nova.scheduler.client.report [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Deleted allocations for instance a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8 [ 967.727921] env[61906]: DEBUG nova.compute.manager [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 967.743114] env[61906]: DEBUG oslo_vmware.api [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333481, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.778858] env[61906]: DEBUG oslo_concurrency.lockutils [None req-46ccdee1-a43e-46ca-9231-9d56829a5e4d tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "bddfdc7f-09d7-4887-b330-34a596ffa562" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.057s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.871990] env[61906]: DEBUG nova.network.neutron [-] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.036090] env[61906]: DEBUG oslo_concurrency.lockutils [None req-42145e43-4e98-4763-ad29-f8292627b20b tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.399s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.036674] env[61906]: DEBUG oslo_concurrency.lockutils [req-363ee9e3-0e04-40ba-829c-f5c682030bc6 req-3c7455fd-92a6-49e8-b1d8-086ed93c471d service nova] Acquired lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.037657] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01589617-7a33-46b2-bfb4-564b07880fb8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.049528] env[61906]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 968.049696] env[61906]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=61906) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 968.050572] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-02b191ca-b14f-4298-a102-c0a14b6feb8d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.059508] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aba2bdf-8304-4846-b591-8f5c3bf0fb0b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.089290] env[61906]: ERROR root [req-363ee9e3-0e04-40ba-829c-f5c682030bc6 req-3c7455fd-92a6-49e8-b1d8-086ed93c471d service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-284752' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-284752' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-284752' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-284752'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-284752' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-284752' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-284752'}\n"]: nova.exception.InstanceNotFound: Instance a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8 could not be found. [ 968.089527] env[61906]: DEBUG oslo_concurrency.lockutils [req-363ee9e3-0e04-40ba-829c-f5c682030bc6 req-3c7455fd-92a6-49e8-b1d8-086ed93c471d service nova] Releasing lock "a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.090180] env[61906]: DEBUG nova.compute.manager [req-363ee9e3-0e04-40ba-829c-f5c682030bc6 req-3c7455fd-92a6-49e8-b1d8-086ed93c471d service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Detach interface failed, port_id=c2232d91-1c2c-4a66-9026-933169da90e1, reason: Instance a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 968.090180] env[61906]: DEBUG nova.compute.manager [req-363ee9e3-0e04-40ba-829c-f5c682030bc6 req-3c7455fd-92a6-49e8-b1d8-086ed93c471d service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Received event network-vif-deleted-4f8ab1c1-fd91-48d3-90e0-ed5ff9fa0b4d {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 968.090180] env[61906]: INFO nova.compute.manager [req-363ee9e3-0e04-40ba-829c-f5c682030bc6 req-3c7455fd-92a6-49e8-b1d8-086ed93c471d service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Neutron deleted interface 4f8ab1c1-fd91-48d3-90e0-ed5ff9fa0b4d; detaching it from the instance and deleting it from the info cache [ 968.090405] env[61906]: DEBUG nova.network.neutron [req-363ee9e3-0e04-40ba-829c-f5c682030bc6 req-3c7455fd-92a6-49e8-b1d8-086ed93c471d service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Updating instance_info_cache with network_info: [{"id": "3e04c5c7-8d7a-46ae-a2d7-52dc429650f9", "address": "fa:16:3e:f1:15:76", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e04c5c7-8d", "ovs_interfaceid": "3e04c5c7-8d7a-46ae-a2d7-52dc429650f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.245593] env[61906]: DEBUG oslo_vmware.api [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333481, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.252829] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.375164] env[61906]: INFO nova.compute.manager [-] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Took 1.02 seconds to deallocate network for instance. [ 968.479851] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "7142e219-6150-4c6f-9632-cbf489035431" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.480270] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "7142e219-6150-4c6f-9632-cbf489035431" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.480578] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "7142e219-6150-4c6f-9632-cbf489035431-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.480881] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "7142e219-6150-4c6f-9632-cbf489035431-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.481171] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "7142e219-6150-4c6f-9632-cbf489035431-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.484136] env[61906]: INFO nova.compute.manager [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Terminating instance [ 968.486608] env[61906]: DEBUG nova.compute.manager [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 968.486941] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 968.488133] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41c71f25-1484-450e-8dee-77d5b64bd3f8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.498287] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 968.498725] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2d4bb9e0-30f1-4ff6-9458-aee8d75b384f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.506150] env[61906]: DEBUG oslo_vmware.api [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 968.506150] env[61906]: value = "task-1333482" [ 968.506150] env[61906]: _type = "Task" [ 968.506150] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.516894] env[61906]: DEBUG oslo_vmware.api [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333482, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.545448] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "bddfdc7f-09d7-4887-b330-34a596ffa562" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.545831] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "bddfdc7f-09d7-4887-b330-34a596ffa562" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.546152] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "bddfdc7f-09d7-4887-b330-34a596ffa562-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.546459] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "bddfdc7f-09d7-4887-b330-34a596ffa562-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.546767] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "bddfdc7f-09d7-4887-b330-34a596ffa562-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.549843] env[61906]: INFO nova.compute.manager [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Terminating instance [ 968.553043] env[61906]: DEBUG nova.compute.manager [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 968.553634] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 968.554927] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3860cf8d-4e90-47a5-8129-068eabeca631 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.565647] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 968.569490] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-35683adb-59ae-4636-bc7d-86a993b2cde6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.578402] env[61906]: DEBUG oslo_vmware.api [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 968.578402] env[61906]: value = "task-1333483" [ 968.578402] env[61906]: _type = "Task" [ 968.578402] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.589323] env[61906]: DEBUG oslo_vmware.api [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333483, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.598715] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-200522f9-e4b8-41bc-a274-b7ef8d38b8bb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.612218] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43b7e183-1f86-4554-b287-0c57bd6e16be {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.644323] env[61906]: DEBUG nova.compute.manager [req-363ee9e3-0e04-40ba-829c-f5c682030bc6 req-3c7455fd-92a6-49e8-b1d8-086ed93c471d service nova] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Detach interface failed, port_id=4f8ab1c1-fd91-48d3-90e0-ed5ff9fa0b4d, reason: Instance a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 968.747807] env[61906]: DEBUG oslo_vmware.api [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333481, 'name': CreateSnapshot_Task, 'duration_secs': 1.220924} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.748214] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Created Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 968.751677] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-134712b9-06d0-4f30-941f-ca621f29ae8f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.783798] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01d0c68a-52ea-479b-9a86-0f53f9de825a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.791860] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e268ba5d-8f7b-45a1-8a63-006c6c9933e0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.826403] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a9d0fad-8d46-4473-a96e-1b8990c1a39e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.836086] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f28cdcc7-4976-46dd-a323-73da0d10cb3b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.853544] env[61906]: DEBUG nova.compute.provider_tree [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.881618] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.016941] env[61906]: DEBUG oslo_vmware.api [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333482, 'name': PowerOffVM_Task, 'duration_secs': 0.265691} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.017331] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 969.017551] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 969.017863] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5984b0de-0176-4205-8a97-b67e1db202c8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.088437] env[61906]: DEBUG oslo_vmware.api [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333483, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.270989] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Creating linked-clone VM from snapshot {{(pid=61906) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 969.271457] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-fcdb27d9-0ab2-40ff-a403-19fd495bcc50 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.279581] env[61906]: DEBUG oslo_vmware.api [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 969.279581] env[61906]: value = "task-1333485" [ 969.279581] env[61906]: _type = "Task" [ 969.279581] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.288157] env[61906]: DEBUG oslo_vmware.api [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333485, 'name': CloneVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.358111] env[61906]: DEBUG nova.scheduler.client.report [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 969.374236] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 969.374469] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 969.374730] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Deleting the datastore file [datastore1] 7142e219-6150-4c6f-9632-cbf489035431 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 969.376746] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-48fc8d46-c6bb-443d-b0f3-ec9e63fbfa95 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.383581] env[61906]: DEBUG oslo_vmware.api [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 969.383581] env[61906]: value = "task-1333486" [ 969.383581] env[61906]: _type = "Task" [ 969.383581] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.392765] env[61906]: DEBUG oslo_vmware.api [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333486, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.590049] env[61906]: DEBUG oslo_vmware.api [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333483, 'name': PowerOffVM_Task, 'duration_secs': 0.69848} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.590344] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 969.590604] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 969.590909] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-70e7ae30-1cb0-4922-9cdc-7c27d312e79c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.632431] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c77db64b-6fd0-4ce2-bc4e-bc94110066fb tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "1a335019-915b-4065-b22f-d2f0af728d6f" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.632817] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c77db64b-6fd0-4ce2-bc4e-bc94110066fb tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "1a335019-915b-4065-b22f-d2f0af728d6f" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.633029] env[61906]: DEBUG nova.compute.manager [None req-c77db64b-6fd0-4ce2-bc4e-bc94110066fb tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 969.634055] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-085a96d1-eda2-43f2-8c98-183e0d7781a5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.641056] env[61906]: DEBUG nova.compute.manager [None req-c77db64b-6fd0-4ce2-bc4e-bc94110066fb tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61906) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 969.641639] env[61906]: DEBUG nova.objects.instance [None req-c77db64b-6fd0-4ce2-bc4e-bc94110066fb tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lazy-loading 'flavor' on Instance uuid 1a335019-915b-4065-b22f-d2f0af728d6f {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 969.657071] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 969.657071] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 969.657071] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Deleting the datastore file [datastore1] bddfdc7f-09d7-4887-b330-34a596ffa562 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 969.657071] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e9bac02f-8f78-4e61-9bce-4607bc75a313 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.665563] env[61906]: DEBUG oslo_vmware.api [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 969.665563] env[61906]: value = "task-1333488" [ 969.665563] env[61906]: _type = "Task" [ 969.665563] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.673539] env[61906]: DEBUG oslo_vmware.api [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333488, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.794190] env[61906]: DEBUG oslo_vmware.api [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333485, 'name': CloneVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.863337] env[61906]: DEBUG oslo_concurrency.lockutils [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.354s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.863921] env[61906]: DEBUG nova.compute.manager [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 969.867404] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.820s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.868927] env[61906]: INFO nova.compute.claims [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 969.892948] env[61906]: DEBUG oslo_vmware.api [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333486, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.212204} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.893251] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 969.893439] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 969.893874] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 969.893874] env[61906]: INFO nova.compute.manager [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Took 1.41 seconds to destroy the instance on the hypervisor. [ 969.894056] env[61906]: DEBUG oslo.service.loopingcall [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 969.894246] env[61906]: DEBUG nova.compute.manager [-] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 969.894335] env[61906]: DEBUG nova.network.neutron [-] [instance: 7142e219-6150-4c6f-9632-cbf489035431] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 970.148667] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c77db64b-6fd0-4ce2-bc4e-bc94110066fb tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 970.149709] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-57c27c61-1c72-4cf8-8905-19e8a69e8813 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.155948] env[61906]: DEBUG oslo_vmware.api [None req-c77db64b-6fd0-4ce2-bc4e-bc94110066fb tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 970.155948] env[61906]: value = "task-1333489" [ 970.155948] env[61906]: _type = "Task" [ 970.155948] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.170314] env[61906]: DEBUG oslo_vmware.api [None req-c77db64b-6fd0-4ce2-bc4e-bc94110066fb tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333489, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.181129] env[61906]: DEBUG oslo_vmware.api [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333488, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16802} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.181516] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 970.181773] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 970.182079] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 970.182268] env[61906]: INFO nova.compute.manager [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Took 1.63 seconds to destroy the instance on the hypervisor. [ 970.182625] env[61906]: DEBUG oslo.service.loopingcall [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 970.183770] env[61906]: DEBUG nova.compute.manager [-] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 970.183867] env[61906]: DEBUG nova.network.neutron [-] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 970.186839] env[61906]: DEBUG nova.compute.manager [req-1109766b-9503-4f2e-bd3e-ef952b83bca0 req-57851432-4730-438a-8cf8-25e54daf2293 service nova] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Received event network-vif-deleted-ff7dc3ad-711f-4510-8a80-43f9717aeeba {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 970.187094] env[61906]: INFO nova.compute.manager [req-1109766b-9503-4f2e-bd3e-ef952b83bca0 req-57851432-4730-438a-8cf8-25e54daf2293 service nova] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Neutron deleted interface ff7dc3ad-711f-4510-8a80-43f9717aeeba; detaching it from the instance and deleting it from the info cache [ 970.187338] env[61906]: DEBUG nova.network.neutron [req-1109766b-9503-4f2e-bd3e-ef952b83bca0 req-57851432-4730-438a-8cf8-25e54daf2293 service nova] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.297188] env[61906]: DEBUG oslo_vmware.api [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333485, 'name': CloneVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.373915] env[61906]: DEBUG nova.compute.utils [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 970.377779] env[61906]: DEBUG nova.compute.manager [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 970.377996] env[61906]: DEBUG nova.network.neutron [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 970.424771] env[61906]: DEBUG nova.policy [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b49e5a6aeb1d43e290a645e14861b889', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '53644a4143a24759a9ff2b5e28b84fb4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 970.549486] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "7264af54-cd10-4507-8ae3-5bb8a5cd127f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.549486] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "7264af54-cd10-4507-8ae3-5bb8a5cd127f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.664658] env[61906]: DEBUG nova.network.neutron [-] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.676158] env[61906]: DEBUG oslo_vmware.api [None req-c77db64b-6fd0-4ce2-bc4e-bc94110066fb tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333489, 'name': PowerOffVM_Task, 'duration_secs': 0.389113} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.677125] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c77db64b-6fd0-4ce2-bc4e-bc94110066fb tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 970.677359] env[61906]: DEBUG nova.compute.manager [None req-c77db64b-6fd0-4ce2-bc4e-bc94110066fb tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 970.678451] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf03ff49-de2e-4ba3-8857-92933063fa25 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.682500] env[61906]: DEBUG nova.network.neutron [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Successfully created port: a8e2f461-ff1a-4839-a351-a9a1925b7be2 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 970.691445] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-592e02a4-609a-4536-b2c9-e64b00e7dbe8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.698390] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50196394-32ce-433e-be44-fce7fbad51b3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.725281] env[61906]: DEBUG nova.compute.manager [req-1109766b-9503-4f2e-bd3e-ef952b83bca0 req-57851432-4730-438a-8cf8-25e54daf2293 service nova] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Detach interface failed, port_id=ff7dc3ad-711f-4510-8a80-43f9717aeeba, reason: Instance 7142e219-6150-4c6f-9632-cbf489035431 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 970.791822] env[61906]: DEBUG oslo_vmware.api [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333485, 'name': CloneVM_Task, 'duration_secs': 1.46848} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.792119] env[61906]: INFO nova.virt.vmwareapi.vmops [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Created linked-clone VM from snapshot [ 970.792916] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7a1afbb-836f-470a-a86f-4286943afff4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.800807] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Uploading image 40b9307b-1417-46aa-b262-7664cfccd96f {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 970.827442] env[61906]: DEBUG oslo_vmware.rw_handles [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 970.827442] env[61906]: value = "vm-284812" [ 970.827442] env[61906]: _type = "VirtualMachine" [ 970.827442] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 970.827752] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-675141d9-b27b-472e-9df1-c98c2c1c8b6d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.834668] env[61906]: DEBUG oslo_vmware.rw_handles [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lease: (returnval){ [ 970.834668] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5252b9a2-85b1-2501-b9b6-5fd24640d375" [ 970.834668] env[61906]: _type = "HttpNfcLease" [ 970.834668] env[61906]: } obtained for exporting VM: (result){ [ 970.834668] env[61906]: value = "vm-284812" [ 970.834668] env[61906]: _type = "VirtualMachine" [ 970.834668] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 970.834988] env[61906]: DEBUG oslo_vmware.api [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the lease: (returnval){ [ 970.834988] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5252b9a2-85b1-2501-b9b6-5fd24640d375" [ 970.834988] env[61906]: _type = "HttpNfcLease" [ 970.834988] env[61906]: } to be ready. {{(pid=61906) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 970.841659] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 970.841659] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5252b9a2-85b1-2501-b9b6-5fd24640d375" [ 970.841659] env[61906]: _type = "HttpNfcLease" [ 970.841659] env[61906]: } is initializing. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 970.878964] env[61906]: DEBUG nova.compute.manager [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 970.906478] env[61906]: DEBUG nova.network.neutron [-] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.051579] env[61906]: DEBUG nova.compute.manager [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 971.057195] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93af4b23-2f5e-4eaa-aa39-bb457a104463 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.064973] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-100c1d76-d368-4555-8795-7b3c0c9494f5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.096238] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c15761ab-2c1b-43b6-9de8-8eec22e6f361 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.104474] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38343ead-5fc1-4df4-acb0-0e9c63c25e0f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.118237] env[61906]: DEBUG nova.compute.provider_tree [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 971.167093] env[61906]: INFO nova.compute.manager [-] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Took 1.27 seconds to deallocate network for instance. [ 971.195863] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c77db64b-6fd0-4ce2-bc4e-bc94110066fb tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "1a335019-915b-4065-b22f-d2f0af728d6f" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.563s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.344237] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 971.344237] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5252b9a2-85b1-2501-b9b6-5fd24640d375" [ 971.344237] env[61906]: _type = "HttpNfcLease" [ 971.344237] env[61906]: } is ready. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 971.344625] env[61906]: DEBUG oslo_vmware.rw_handles [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 971.344625] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5252b9a2-85b1-2501-b9b6-5fd24640d375" [ 971.344625] env[61906]: _type = "HttpNfcLease" [ 971.344625] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 971.345245] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01f7d0f-d8ab-4d0e-8c74-802bb2ae72a3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.352419] env[61906]: DEBUG oslo_vmware.rw_handles [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ef10c7-d4bc-2624-0865-b110aac993bd/disk-0.vmdk from lease info. {{(pid=61906) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 971.352813] env[61906]: DEBUG oslo_vmware.rw_handles [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ef10c7-d4bc-2624-0865-b110aac993bd/disk-0.vmdk for reading. {{(pid=61906) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 971.415723] env[61906]: INFO nova.compute.manager [-] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Took 1.23 seconds to deallocate network for instance. [ 971.513774] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-da4615ae-2c8f-4d41-a31a-12eb48a9ef5a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.570165] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.641257] env[61906]: ERROR nova.scheduler.client.report [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [req-9445bde1-db12-477f-9533-b50573e26b52] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID a5b57df7-a16e-44f9-9b2d-23c518860263. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-9445bde1-db12-477f-9533-b50573e26b52"}]} [ 971.656864] env[61906]: DEBUG nova.scheduler.client.report [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Refreshing inventories for resource provider a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 971.670890] env[61906]: DEBUG nova.scheduler.client.report [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Updating ProviderTree inventory for provider a5b57df7-a16e-44f9-9b2d-23c518860263 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 971.671159] env[61906]: DEBUG nova.compute.provider_tree [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 971.677063] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.687535] env[61906]: DEBUG nova.scheduler.client.report [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Refreshing aggregate associations for resource provider a5b57df7-a16e-44f9-9b2d-23c518860263, aggregates: None {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 971.706394] env[61906]: DEBUG nova.scheduler.client.report [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Refreshing trait associations for resource provider a5b57df7-a16e-44f9-9b2d-23c518860263, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64 {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 971.900146] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44a9447a-8954-41c3-935f-e2bc1cac19cf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.908881] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dc32431-dc9d-4f11-96a3-88a7fad4d853 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.943363] env[61906]: DEBUG nova.compute.manager [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 971.946472] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.947932] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbe56c95-1c4f-406e-81f8-dea4f5a9d095 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.956520] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62d81198-fb0e-4deb-bf3e-7be8ab8c5735 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.974934] env[61906]: DEBUG nova.compute.provider_tree [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 971.979822] env[61906]: DEBUG nova.virt.hardware [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 971.980105] env[61906]: DEBUG nova.virt.hardware [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 971.980275] env[61906]: DEBUG nova.virt.hardware [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 971.980463] env[61906]: DEBUG nova.virt.hardware [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 971.980616] env[61906]: DEBUG nova.virt.hardware [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 971.980771] env[61906]: DEBUG nova.virt.hardware [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 971.980980] env[61906]: DEBUG nova.virt.hardware [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 971.981160] env[61906]: DEBUG nova.virt.hardware [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 971.981331] env[61906]: DEBUG nova.virt.hardware [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 971.981521] env[61906]: DEBUG nova.virt.hardware [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 971.981753] env[61906]: DEBUG nova.virt.hardware [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 971.982867] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55ae59a4-b019-4974-8b4e-82c05d01aaf2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.992715] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f23136f-565f-4d5f-a0b6-b1ec8da7fde6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.025947] env[61906]: DEBUG nova.objects.instance [None req-7e10905f-2adb-4157-ba38-1bc867eb4262 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lazy-loading 'flavor' on Instance uuid 1a335019-915b-4065-b22f-d2f0af728d6f {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 972.145118] env[61906]: DEBUG nova.compute.manager [req-45db16a2-07d5-44b5-824d-9ca50699a99f req-4c79bac1-d9e3-407d-8c7a-3f2a7d9d352c service nova] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Received event network-vif-plugged-a8e2f461-ff1a-4839-a351-a9a1925b7be2 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 972.145118] env[61906]: DEBUG oslo_concurrency.lockutils [req-45db16a2-07d5-44b5-824d-9ca50699a99f req-4c79bac1-d9e3-407d-8c7a-3f2a7d9d352c service nova] Acquiring lock "f49fb403-fbb4-4e26-8e75-c160d11dea05-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.145118] env[61906]: DEBUG oslo_concurrency.lockutils [req-45db16a2-07d5-44b5-824d-9ca50699a99f req-4c79bac1-d9e3-407d-8c7a-3f2a7d9d352c service nova] Lock "f49fb403-fbb4-4e26-8e75-c160d11dea05-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.145118] env[61906]: DEBUG oslo_concurrency.lockutils [req-45db16a2-07d5-44b5-824d-9ca50699a99f req-4c79bac1-d9e3-407d-8c7a-3f2a7d9d352c service nova] Lock "f49fb403-fbb4-4e26-8e75-c160d11dea05-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.145118] env[61906]: DEBUG nova.compute.manager [req-45db16a2-07d5-44b5-824d-9ca50699a99f req-4c79bac1-d9e3-407d-8c7a-3f2a7d9d352c service nova] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] No waiting events found dispatching network-vif-plugged-a8e2f461-ff1a-4839-a351-a9a1925b7be2 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 972.145118] env[61906]: WARNING nova.compute.manager [req-45db16a2-07d5-44b5-824d-9ca50699a99f req-4c79bac1-d9e3-407d-8c7a-3f2a7d9d352c service nova] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Received unexpected event network-vif-plugged-a8e2f461-ff1a-4839-a351-a9a1925b7be2 for instance with vm_state building and task_state spawning. [ 972.224829] env[61906]: DEBUG nova.compute.manager [req-d2d16bfa-2909-4f4a-a144-dd70a03796fe req-204a177e-b972-494a-9935-a4c55a2a4c0c service nova] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Received event network-vif-deleted-c45bb207-8a87-4c79-8247-704fd5fd810b {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 972.277677] env[61906]: DEBUG nova.network.neutron [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Successfully updated port: a8e2f461-ff1a-4839-a351-a9a1925b7be2 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 972.516503] env[61906]: DEBUG nova.scheduler.client.report [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Updated inventory for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with generation 103 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 972.516953] env[61906]: DEBUG nova.compute.provider_tree [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Updating resource provider a5b57df7-a16e-44f9-9b2d-23c518860263 generation from 103 to 104 during operation: update_inventory {{(pid=61906) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 972.518627] env[61906]: DEBUG nova.compute.provider_tree [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 972.534255] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7e10905f-2adb-4157-ba38-1bc867eb4262 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.534750] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7e10905f-2adb-4157-ba38-1bc867eb4262 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.535113] env[61906]: DEBUG nova.network.neutron [None req-7e10905f-2adb-4157-ba38-1bc867eb4262 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 972.535861] env[61906]: DEBUG nova.objects.instance [None req-7e10905f-2adb-4157-ba38-1bc867eb4262 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lazy-loading 'info_cache' on Instance uuid 1a335019-915b-4065-b22f-d2f0af728d6f {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 972.785699] env[61906]: DEBUG oslo_concurrency.lockutils [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "refresh_cache-f49fb403-fbb4-4e26-8e75-c160d11dea05" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.785699] env[61906]: DEBUG oslo_concurrency.lockutils [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired lock "refresh_cache-f49fb403-fbb4-4e26-8e75-c160d11dea05" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.785699] env[61906]: DEBUG nova.network.neutron [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 973.022408] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.155s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.024120] env[61906]: DEBUG nova.compute.manager [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 973.025648] env[61906]: DEBUG oslo_concurrency.lockutils [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.162s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.025868] env[61906]: DEBUG oslo_concurrency.lockutils [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.027942] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.775s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.029420] env[61906]: INFO nova.compute.claims [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 973.038865] env[61906]: DEBUG nova.objects.base [None req-7e10905f-2adb-4157-ba38-1bc867eb4262 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Object Instance<1a335019-915b-4065-b22f-d2f0af728d6f> lazy-loaded attributes: flavor,info_cache {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 973.051213] env[61906]: INFO nova.scheduler.client.report [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Deleted allocations for instance e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15 [ 973.317365] env[61906]: DEBUG nova.network.neutron [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 973.511403] env[61906]: DEBUG nova.network.neutron [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Updating instance_info_cache with network_info: [{"id": "a8e2f461-ff1a-4839-a351-a9a1925b7be2", "address": "fa:16:3e:8b:b6:8e", "network": {"id": "bc9185e9-f26b-489c-84ee-e6db41ac6283", "bridge": "br-int", "label": "tempest-ServersTestJSON-1167747148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53644a4143a24759a9ff2b5e28b84fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8e2f461-ff", "ovs_interfaceid": "a8e2f461-ff1a-4839-a351-a9a1925b7be2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.534637] env[61906]: DEBUG nova.compute.utils [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 973.537749] env[61906]: DEBUG nova.compute.manager [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 973.537955] env[61906]: DEBUG nova.network.neutron [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 973.559929] env[61906]: DEBUG oslo_concurrency.lockutils [None req-89ee660d-9499-4898-8b2d-503e67a7c8c2 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.120s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.581472] env[61906]: DEBUG nova.policy [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bfb8ebd8f94f45dfadb1d8802d04aec9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4f7fc2042e6549cabe98838990f35842', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 973.764979] env[61906]: DEBUG nova.network.neutron [None req-7e10905f-2adb-4157-ba38-1bc867eb4262 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Updating instance_info_cache with network_info: [{"id": "a497b774-60ea-485d-a564-ba7b978b560e", "address": "fa:16:3e:0d:cf:6e", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa497b774-60", "ovs_interfaceid": "a497b774-60ea-485d-a564-ba7b978b560e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.850122] env[61906]: DEBUG nova.network.neutron [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Successfully created port: da64f0e8-39b7-4d88-aa6d-4ac0939a0a20 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 974.013964] env[61906]: DEBUG oslo_concurrency.lockutils [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Releasing lock "refresh_cache-f49fb403-fbb4-4e26-8e75-c160d11dea05" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.014421] env[61906]: DEBUG nova.compute.manager [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Instance network_info: |[{"id": "a8e2f461-ff1a-4839-a351-a9a1925b7be2", "address": "fa:16:3e:8b:b6:8e", "network": {"id": "bc9185e9-f26b-489c-84ee-e6db41ac6283", "bridge": "br-int", "label": "tempest-ServersTestJSON-1167747148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53644a4143a24759a9ff2b5e28b84fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8e2f461-ff", "ovs_interfaceid": "a8e2f461-ff1a-4839-a351-a9a1925b7be2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 974.014773] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8b:b6:8e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a8e2f461-ff1a-4839-a351-a9a1925b7be2', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 974.022313] env[61906]: DEBUG oslo.service.loopingcall [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 974.022531] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 974.022789] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e1768f17-7c5d-4b7d-91d2-e04e95e5fea8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.041556] env[61906]: DEBUG nova.compute.manager [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 974.049021] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 974.049021] env[61906]: value = "task-1333491" [ 974.049021] env[61906]: _type = "Task" [ 974.049021] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.057104] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333491, 'name': CreateVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.217450] env[61906]: DEBUG nova.compute.manager [req-00794207-bcdb-47ca-83e9-d94cee3c0897 req-165b69bf-1e97-42bd-9620-74966d16e1d1 service nova] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Received event network-changed-a8e2f461-ff1a-4839-a351-a9a1925b7be2 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 974.217789] env[61906]: DEBUG nova.compute.manager [req-00794207-bcdb-47ca-83e9-d94cee3c0897 req-165b69bf-1e97-42bd-9620-74966d16e1d1 service nova] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Refreshing instance network info cache due to event network-changed-a8e2f461-ff1a-4839-a351-a9a1925b7be2. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 974.218155] env[61906]: DEBUG oslo_concurrency.lockutils [req-00794207-bcdb-47ca-83e9-d94cee3c0897 req-165b69bf-1e97-42bd-9620-74966d16e1d1 service nova] Acquiring lock "refresh_cache-f49fb403-fbb4-4e26-8e75-c160d11dea05" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.218416] env[61906]: DEBUG oslo_concurrency.lockutils [req-00794207-bcdb-47ca-83e9-d94cee3c0897 req-165b69bf-1e97-42bd-9620-74966d16e1d1 service nova] Acquired lock "refresh_cache-f49fb403-fbb4-4e26-8e75-c160d11dea05" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.218700] env[61906]: DEBUG nova.network.neutron [req-00794207-bcdb-47ca-83e9-d94cee3c0897 req-165b69bf-1e97-42bd-9620-74966d16e1d1 service nova] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Refreshing network info cache for port a8e2f461-ff1a-4839-a351-a9a1925b7be2 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 974.251315] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-849fb936-c691-4ff4-bb99-5fcb80f18687 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.259249] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-867837f3-6f56-4e35-bda0-b9e96908ad4d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.267094] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7e10905f-2adb-4157-ba38-1bc867eb4262 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Releasing lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.294692] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cedd3b59-4e1a-4416-b726-5db84d19bba8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.303114] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf98ded2-5b17-4d08-9b15-4fb904e0337c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.317204] env[61906]: DEBUG nova.compute.provider_tree [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 974.560950] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333491, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.795470] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e10905f-2adb-4157-ba38-1bc867eb4262 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 974.795804] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0317c554-3b56-4baf-a57d-be06ad319a57 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.805015] env[61906]: DEBUG oslo_vmware.api [None req-7e10905f-2adb-4157-ba38-1bc867eb4262 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 974.805015] env[61906]: value = "task-1333492" [ 974.805015] env[61906]: _type = "Task" [ 974.805015] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.812839] env[61906]: DEBUG oslo_vmware.api [None req-7e10905f-2adb-4157-ba38-1bc867eb4262 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333492, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.820724] env[61906]: DEBUG nova.scheduler.client.report [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 974.937608] env[61906]: DEBUG nova.network.neutron [req-00794207-bcdb-47ca-83e9-d94cee3c0897 req-165b69bf-1e97-42bd-9620-74966d16e1d1 service nova] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Updated VIF entry in instance network info cache for port a8e2f461-ff1a-4839-a351-a9a1925b7be2. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 974.938017] env[61906]: DEBUG nova.network.neutron [req-00794207-bcdb-47ca-83e9-d94cee3c0897 req-165b69bf-1e97-42bd-9620-74966d16e1d1 service nova] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Updating instance_info_cache with network_info: [{"id": "a8e2f461-ff1a-4839-a351-a9a1925b7be2", "address": "fa:16:3e:8b:b6:8e", "network": {"id": "bc9185e9-f26b-489c-84ee-e6db41ac6283", "bridge": "br-int", "label": "tempest-ServersTestJSON-1167747148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53644a4143a24759a9ff2b5e28b84fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8e2f461-ff", "ovs_interfaceid": "a8e2f461-ff1a-4839-a351-a9a1925b7be2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.058038] env[61906]: DEBUG nova.compute.manager [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 975.063688] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333491, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.083373] env[61906]: DEBUG nova.virt.hardware [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 975.083615] env[61906]: DEBUG nova.virt.hardware [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 975.083777] env[61906]: DEBUG nova.virt.hardware [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 975.083966] env[61906]: DEBUG nova.virt.hardware [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 975.084124] env[61906]: DEBUG nova.virt.hardware [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 975.084744] env[61906]: DEBUG nova.virt.hardware [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 975.084744] env[61906]: DEBUG nova.virt.hardware [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 975.084744] env[61906]: DEBUG nova.virt.hardware [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 975.084942] env[61906]: DEBUG nova.virt.hardware [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 975.084942] env[61906]: DEBUG nova.virt.hardware [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 975.085114] env[61906]: DEBUG nova.virt.hardware [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 975.085948] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a65d5d-e4a1-4c78-9215-b0923d9a5410 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.093699] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78ab3d0c-37b8-4f86-b189-0384d18266a0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.315616] env[61906]: DEBUG oslo_vmware.api [None req-7e10905f-2adb-4157-ba38-1bc867eb4262 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333492, 'name': PowerOnVM_Task, 'duration_secs': 0.458531} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.315943] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e10905f-2adb-4157-ba38-1bc867eb4262 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 975.316177] env[61906]: DEBUG nova.compute.manager [None req-7e10905f-2adb-4157-ba38-1bc867eb4262 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 975.317112] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52007901-7086-4db7-8c77-8cf8ac2378f4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.325902] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.298s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.326400] env[61906]: DEBUG nova.compute.manager [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 975.328647] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.447s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.328864] env[61906]: DEBUG nova.objects.instance [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Lazy-loading 'resources' on Instance uuid 11cb9d89-2304-45a6-b4bb-fcf8417a0518 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 975.430030] env[61906]: DEBUG nova.network.neutron [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Successfully updated port: da64f0e8-39b7-4d88-aa6d-4ac0939a0a20 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 975.440143] env[61906]: DEBUG oslo_concurrency.lockutils [req-00794207-bcdb-47ca-83e9-d94cee3c0897 req-165b69bf-1e97-42bd-9620-74966d16e1d1 service nova] Releasing lock "refresh_cache-f49fb403-fbb4-4e26-8e75-c160d11dea05" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.562256] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333491, 'name': CreateVM_Task, 'duration_secs': 1.389064} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.562493] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 975.563314] env[61906]: DEBUG oslo_concurrency.lockutils [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.563501] env[61906]: DEBUG oslo_concurrency.lockutils [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.563894] env[61906]: DEBUG oslo_concurrency.lockutils [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 975.564211] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f737cd2c-1e94-45b3-8e00-321b42854b30 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.569161] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 975.569161] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52ef728c-8408-2d71-bed6-991232e62f29" [ 975.569161] env[61906]: _type = "Task" [ 975.569161] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.577337] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52ef728c-8408-2d71-bed6-991232e62f29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.832076] env[61906]: DEBUG nova.compute.utils [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 975.832698] env[61906]: DEBUG nova.compute.manager [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 975.833082] env[61906]: DEBUG nova.network.neutron [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 975.876986] env[61906]: DEBUG nova.policy [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e89d45df95134c709b569d06744dea5e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1b33dea0794a48f78b9f519cb269a8c2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 975.931959] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "refresh_cache-458a1669-a62b-4313-874a-e49809d5c034" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.932137] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired lock "refresh_cache-458a1669-a62b-4313-874a-e49809d5c034" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.932288] env[61906]: DEBUG nova.network.neutron [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 975.997989] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18c80a10-94f3-4554-ae01-e5bd594479a7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.005565] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1236937-73ca-4331-9360-cf55fbbf170b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.037331] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be0d15dc-809f-4d01-873a-5e4cc653dc55 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.044980] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-480bbc53-9600-4fc7-abc3-62326799f2ff {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.059081] env[61906]: DEBUG nova.compute.provider_tree [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 976.079574] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52ef728c-8408-2d71-bed6-991232e62f29, 'name': SearchDatastore_Task, 'duration_secs': 0.037576} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.079903] env[61906]: DEBUG oslo_concurrency.lockutils [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.080216] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 976.080510] env[61906]: DEBUG oslo_concurrency.lockutils [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.080654] env[61906]: DEBUG oslo_concurrency.lockutils [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.080860] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 976.081247] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-080a2536-b974-4f8d-b9c2-db8d716e29eb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.090265] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 976.090740] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 976.091209] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a45cd660-ace5-49aa-a526-63ef22011cdc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.096542] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 976.096542] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5279fadb-c2a6-a356-186c-4ac4f1afe707" [ 976.096542] env[61906]: _type = "Task" [ 976.096542] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.104502] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5279fadb-c2a6-a356-186c-4ac4f1afe707, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.129041] env[61906]: DEBUG nova.network.neutron [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Successfully created port: e3ce81bf-b27e-47ea-9c48-a17e608a9cd6 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 976.334845] env[61906]: DEBUG nova.compute.manager [req-57951a40-9180-4a8b-927b-209b43db14c6 req-07238946-fb7c-4ec0-8e10-6e3f158a445b service nova] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Received event network-vif-plugged-da64f0e8-39b7-4d88-aa6d-4ac0939a0a20 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 976.335223] env[61906]: DEBUG oslo_concurrency.lockutils [req-57951a40-9180-4a8b-927b-209b43db14c6 req-07238946-fb7c-4ec0-8e10-6e3f158a445b service nova] Acquiring lock "458a1669-a62b-4313-874a-e49809d5c034-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.335558] env[61906]: DEBUG oslo_concurrency.lockutils [req-57951a40-9180-4a8b-927b-209b43db14c6 req-07238946-fb7c-4ec0-8e10-6e3f158a445b service nova] Lock "458a1669-a62b-4313-874a-e49809d5c034-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.335862] env[61906]: DEBUG oslo_concurrency.lockutils [req-57951a40-9180-4a8b-927b-209b43db14c6 req-07238946-fb7c-4ec0-8e10-6e3f158a445b service nova] Lock "458a1669-a62b-4313-874a-e49809d5c034-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.336101] env[61906]: DEBUG nova.compute.manager [req-57951a40-9180-4a8b-927b-209b43db14c6 req-07238946-fb7c-4ec0-8e10-6e3f158a445b service nova] [instance: 458a1669-a62b-4313-874a-e49809d5c034] No waiting events found dispatching network-vif-plugged-da64f0e8-39b7-4d88-aa6d-4ac0939a0a20 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 976.336371] env[61906]: WARNING nova.compute.manager [req-57951a40-9180-4a8b-927b-209b43db14c6 req-07238946-fb7c-4ec0-8e10-6e3f158a445b service nova] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Received unexpected event network-vif-plugged-da64f0e8-39b7-4d88-aa6d-4ac0939a0a20 for instance with vm_state building and task_state spawning. [ 976.336556] env[61906]: DEBUG nova.compute.manager [req-57951a40-9180-4a8b-927b-209b43db14c6 req-07238946-fb7c-4ec0-8e10-6e3f158a445b service nova] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Received event network-changed-da64f0e8-39b7-4d88-aa6d-4ac0939a0a20 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 976.336804] env[61906]: DEBUG nova.compute.manager [req-57951a40-9180-4a8b-927b-209b43db14c6 req-07238946-fb7c-4ec0-8e10-6e3f158a445b service nova] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Refreshing instance network info cache due to event network-changed-da64f0e8-39b7-4d88-aa6d-4ac0939a0a20. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 976.337096] env[61906]: DEBUG oslo_concurrency.lockutils [req-57951a40-9180-4a8b-927b-209b43db14c6 req-07238946-fb7c-4ec0-8e10-6e3f158a445b service nova] Acquiring lock "refresh_cache-458a1669-a62b-4313-874a-e49809d5c034" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.338196] env[61906]: DEBUG nova.compute.manager [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 976.465594] env[61906]: DEBUG nova.network.neutron [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 976.564499] env[61906]: DEBUG nova.scheduler.client.report [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 976.612739] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5279fadb-c2a6-a356-186c-4ac4f1afe707, 'name': SearchDatastore_Task, 'duration_secs': 0.008504} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.613926] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83f4b824-66fa-40d0-9ce1-cc4753121094 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.623541] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 976.623541] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52ba4ccb-7ed6-5d02-9d56-679eabc65e80" [ 976.623541] env[61906]: _type = "Task" [ 976.623541] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.629587] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52ba4ccb-7ed6-5d02-9d56-679eabc65e80, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.672605] env[61906]: DEBUG nova.network.neutron [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Updating instance_info_cache with network_info: [{"id": "da64f0e8-39b7-4d88-aa6d-4ac0939a0a20", "address": "fa:16:3e:3b:f8:9b", "network": {"id": "af039bdd-ffd2-483d-9655-e7df48936290", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-111531540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f7fc2042e6549cabe98838990f35842", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda64f0e8-39", "ovs_interfaceid": "da64f0e8-39b7-4d88-aa6d-4ac0939a0a20", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.070024] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.741s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.072515] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.503s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.074091] env[61906]: INFO nova.compute.claims [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 977.092410] env[61906]: INFO nova.scheduler.client.report [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Deleted allocations for instance 11cb9d89-2304-45a6-b4bb-fcf8417a0518 [ 977.133315] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52ba4ccb-7ed6-5d02-9d56-679eabc65e80, 'name': SearchDatastore_Task, 'duration_secs': 0.009391} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.134580] env[61906]: DEBUG oslo_concurrency.lockutils [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.134580] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] f49fb403-fbb4-4e26-8e75-c160d11dea05/f49fb403-fbb4-4e26-8e75-c160d11dea05.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 977.134795] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-21d3b6cc-eb7d-4774-af99-d3216c3fcbeb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.143436] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 977.143436] env[61906]: value = "task-1333493" [ 977.143436] env[61906]: _type = "Task" [ 977.143436] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.152443] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333493, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.176124] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Releasing lock "refresh_cache-458a1669-a62b-4313-874a-e49809d5c034" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.176481] env[61906]: DEBUG nova.compute.manager [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Instance network_info: |[{"id": "da64f0e8-39b7-4d88-aa6d-4ac0939a0a20", "address": "fa:16:3e:3b:f8:9b", "network": {"id": "af039bdd-ffd2-483d-9655-e7df48936290", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-111531540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f7fc2042e6549cabe98838990f35842", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda64f0e8-39", "ovs_interfaceid": "da64f0e8-39b7-4d88-aa6d-4ac0939a0a20", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 977.176827] env[61906]: DEBUG oslo_concurrency.lockutils [req-57951a40-9180-4a8b-927b-209b43db14c6 req-07238946-fb7c-4ec0-8e10-6e3f158a445b service nova] Acquired lock "refresh_cache-458a1669-a62b-4313-874a-e49809d5c034" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.177084] env[61906]: DEBUG nova.network.neutron [req-57951a40-9180-4a8b-927b-209b43db14c6 req-07238946-fb7c-4ec0-8e10-6e3f158a445b service nova] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Refreshing network info cache for port da64f0e8-39b7-4d88-aa6d-4ac0939a0a20 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 977.178374] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3b:f8:9b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '19440099-773e-4a31-b82e-84a4daa5d8fe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'da64f0e8-39b7-4d88-aa6d-4ac0939a0a20', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 977.187592] env[61906]: DEBUG oslo.service.loopingcall [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 977.191109] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 977.191673] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a0d34205-fdf4-43d1-b5ff-fec6939d33a7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.212973] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 977.212973] env[61906]: value = "task-1333494" [ 977.212973] env[61906]: _type = "Task" [ 977.212973] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.222289] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333494, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.350091] env[61906]: DEBUG nova.compute.manager [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 977.375824] env[61906]: DEBUG nova.virt.hardware [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 977.376109] env[61906]: DEBUG nova.virt.hardware [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 977.376277] env[61906]: DEBUG nova.virt.hardware [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 977.376476] env[61906]: DEBUG nova.virt.hardware [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 977.376633] env[61906]: DEBUG nova.virt.hardware [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 977.376787] env[61906]: DEBUG nova.virt.hardware [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 977.377013] env[61906]: DEBUG nova.virt.hardware [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 977.377212] env[61906]: DEBUG nova.virt.hardware [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 977.377359] env[61906]: DEBUG nova.virt.hardware [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 977.377560] env[61906]: DEBUG nova.virt.hardware [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 977.377706] env[61906]: DEBUG nova.virt.hardware [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 977.378702] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e6b54fa-06da-4ada-ae80-758a571b43cb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.390438] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3387d84c-98bf-494b-b36b-ed2bcc098423 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.466359] env[61906]: DEBUG nova.network.neutron [req-57951a40-9180-4a8b-927b-209b43db14c6 req-07238946-fb7c-4ec0-8e10-6e3f158a445b service nova] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Updated VIF entry in instance network info cache for port da64f0e8-39b7-4d88-aa6d-4ac0939a0a20. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 977.466793] env[61906]: DEBUG nova.network.neutron [req-57951a40-9180-4a8b-927b-209b43db14c6 req-07238946-fb7c-4ec0-8e10-6e3f158a445b service nova] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Updating instance_info_cache with network_info: [{"id": "da64f0e8-39b7-4d88-aa6d-4ac0939a0a20", "address": "fa:16:3e:3b:f8:9b", "network": {"id": "af039bdd-ffd2-483d-9655-e7df48936290", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-111531540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f7fc2042e6549cabe98838990f35842", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda64f0e8-39", "ovs_interfaceid": "da64f0e8-39b7-4d88-aa6d-4ac0939a0a20", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.602109] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2d536868-3aef-4512-b36d-863779506573 tempest-ServerShowV247Test-1578955311 tempest-ServerShowV247Test-1578955311-project-member] Lock "11cb9d89-2304-45a6-b4bb-fcf8417a0518" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.426s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.657362] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333493, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.728471] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333494, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.967573] env[61906]: DEBUG nova.network.neutron [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Successfully updated port: e3ce81bf-b27e-47ea-9c48-a17e608a9cd6 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 977.971639] env[61906]: DEBUG oslo_concurrency.lockutils [req-57951a40-9180-4a8b-927b-209b43db14c6 req-07238946-fb7c-4ec0-8e10-6e3f158a445b service nova] Releasing lock "refresh_cache-458a1669-a62b-4313-874a-e49809d5c034" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.162490] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333493, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.826307} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.162790] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] f49fb403-fbb4-4e26-8e75-c160d11dea05/f49fb403-fbb4-4e26-8e75-c160d11dea05.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 978.163110] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 978.163287] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-04f04ef7-b55a-47b1-ab31-846ebfec3d59 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.171696] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 978.171696] env[61906]: value = "task-1333495" [ 978.171696] env[61906]: _type = "Task" [ 978.171696] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.181011] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333495, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.224277] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333494, 'name': CreateVM_Task, 'duration_secs': 0.524857} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.224277] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 978.224585] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.224798] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.225147] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 978.227760] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9814ab69-29f8-45cb-a80f-36e29bae4444 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.233061] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 978.233061] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]529792f2-52f7-39fd-b85c-710cb945befb" [ 978.233061] env[61906]: _type = "Task" [ 978.233061] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.241348] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]529792f2-52f7-39fd-b85c-710cb945befb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.255695] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86c0cd1b-ad9c-4408-8236-6f082c3884cd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.262790] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-795c5edb-e17e-44fa-b19f-74873ffd074d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.292777] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb706ef0-ab99-4a9f-82e8-a23ec6cb62cb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.300659] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc17ce4c-a1d9-463f-8e60-4e4f52d3b518 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.314107] env[61906]: DEBUG nova.compute.provider_tree [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 978.366470] env[61906]: DEBUG nova.compute.manager [req-8ca1ecab-b9dd-4059-9df8-9f34609e10f7 req-87bebd31-f611-47ac-8862-f66bd54c00fe service nova] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Received event network-vif-plugged-e3ce81bf-b27e-47ea-9c48-a17e608a9cd6 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 978.366702] env[61906]: DEBUG oslo_concurrency.lockutils [req-8ca1ecab-b9dd-4059-9df8-9f34609e10f7 req-87bebd31-f611-47ac-8862-f66bd54c00fe service nova] Acquiring lock "a3e3ad00-6921-4072-8cb1-d80302883513-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.366917] env[61906]: DEBUG oslo_concurrency.lockutils [req-8ca1ecab-b9dd-4059-9df8-9f34609e10f7 req-87bebd31-f611-47ac-8862-f66bd54c00fe service nova] Lock "a3e3ad00-6921-4072-8cb1-d80302883513-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.367105] env[61906]: DEBUG oslo_concurrency.lockutils [req-8ca1ecab-b9dd-4059-9df8-9f34609e10f7 req-87bebd31-f611-47ac-8862-f66bd54c00fe service nova] Lock "a3e3ad00-6921-4072-8cb1-d80302883513-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.367294] env[61906]: DEBUG nova.compute.manager [req-8ca1ecab-b9dd-4059-9df8-9f34609e10f7 req-87bebd31-f611-47ac-8862-f66bd54c00fe service nova] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] No waiting events found dispatching network-vif-plugged-e3ce81bf-b27e-47ea-9c48-a17e608a9cd6 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 978.367504] env[61906]: WARNING nova.compute.manager [req-8ca1ecab-b9dd-4059-9df8-9f34609e10f7 req-87bebd31-f611-47ac-8862-f66bd54c00fe service nova] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Received unexpected event network-vif-plugged-e3ce81bf-b27e-47ea-9c48-a17e608a9cd6 for instance with vm_state building and task_state spawning. [ 978.367587] env[61906]: DEBUG nova.compute.manager [req-8ca1ecab-b9dd-4059-9df8-9f34609e10f7 req-87bebd31-f611-47ac-8862-f66bd54c00fe service nova] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Received event network-changed-e3ce81bf-b27e-47ea-9c48-a17e608a9cd6 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 978.367741] env[61906]: DEBUG nova.compute.manager [req-8ca1ecab-b9dd-4059-9df8-9f34609e10f7 req-87bebd31-f611-47ac-8862-f66bd54c00fe service nova] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Refreshing instance network info cache due to event network-changed-e3ce81bf-b27e-47ea-9c48-a17e608a9cd6. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 978.367982] env[61906]: DEBUG oslo_concurrency.lockutils [req-8ca1ecab-b9dd-4059-9df8-9f34609e10f7 req-87bebd31-f611-47ac-8862-f66bd54c00fe service nova] Acquiring lock "refresh_cache-a3e3ad00-6921-4072-8cb1-d80302883513" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.368155] env[61906]: DEBUG oslo_concurrency.lockutils [req-8ca1ecab-b9dd-4059-9df8-9f34609e10f7 req-87bebd31-f611-47ac-8862-f66bd54c00fe service nova] Acquired lock "refresh_cache-a3e3ad00-6921-4072-8cb1-d80302883513" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.368320] env[61906]: DEBUG nova.network.neutron [req-8ca1ecab-b9dd-4059-9df8-9f34609e10f7 req-87bebd31-f611-47ac-8862-f66bd54c00fe service nova] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Refreshing network info cache for port e3ce81bf-b27e-47ea-9c48-a17e608a9cd6 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 978.471384] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "refresh_cache-a3e3ad00-6921-4072-8cb1-d80302883513" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.681741] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333495, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077514} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.682128] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 978.682823] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03b33ac5-93bb-4c7f-925c-391617e0f6fc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.704807] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] f49fb403-fbb4-4e26-8e75-c160d11dea05/f49fb403-fbb4-4e26-8e75-c160d11dea05.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 978.705025] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a032f4d-7469-4cad-a5a8-d3da10463b1d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.724569] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 978.724569] env[61906]: value = "task-1333496" [ 978.724569] env[61906]: _type = "Task" [ 978.724569] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.732479] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333496, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.742390] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]529792f2-52f7-39fd-b85c-710cb945befb, 'name': SearchDatastore_Task, 'duration_secs': 0.009474} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.742781] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.743229] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 978.743371] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.743533] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.743717] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 978.744013] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e14ffe5a-99a9-4f0b-bc05-bb3ea5d5a360 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.845872] env[61906]: DEBUG nova.scheduler.client.report [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Updated inventory for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with generation 104 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 978.846225] env[61906]: DEBUG nova.compute.provider_tree [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Updating resource provider a5b57df7-a16e-44f9-9b2d-23c518860263 generation from 104 to 105 during operation: update_inventory {{(pid=61906) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 978.846441] env[61906]: DEBUG nova.compute.provider_tree [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 978.910886] env[61906]: DEBUG nova.network.neutron [req-8ca1ecab-b9dd-4059-9df8-9f34609e10f7 req-87bebd31-f611-47ac-8862-f66bd54c00fe service nova] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 978.989853] env[61906]: DEBUG nova.network.neutron [req-8ca1ecab-b9dd-4059-9df8-9f34609e10f7 req-87bebd31-f611-47ac-8862-f66bd54c00fe service nova] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.234860] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333496, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.351555] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.279s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.352193] env[61906]: DEBUG nova.compute.manager [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 979.355029] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.678s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.355281] env[61906]: DEBUG nova.objects.instance [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lazy-loading 'resources' on Instance uuid 7142e219-6150-4c6f-9632-cbf489035431 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 979.493063] env[61906]: DEBUG oslo_concurrency.lockutils [req-8ca1ecab-b9dd-4059-9df8-9f34609e10f7 req-87bebd31-f611-47ac-8862-f66bd54c00fe service nova] Releasing lock "refresh_cache-a3e3ad00-6921-4072-8cb1-d80302883513" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.493468] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquired lock "refresh_cache-a3e3ad00-6921-4072-8cb1-d80302883513" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.493691] env[61906]: DEBUG nova.network.neutron [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 979.737925] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333496, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.792431] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 979.792711] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 979.793723] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6aa549c-a3c3-4875-bcac-bb968de171c0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.800050] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 979.800050] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52bc96c5-ca66-a38e-882f-3112efc372c4" [ 979.800050] env[61906]: _type = "Task" [ 979.800050] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.806892] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52bc96c5-ca66-a38e-882f-3112efc372c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.858824] env[61906]: DEBUG nova.compute.utils [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 979.860327] env[61906]: DEBUG nova.compute.manager [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 979.860493] env[61906]: DEBUG nova.network.neutron [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 979.902761] env[61906]: DEBUG nova.policy [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0de9abfd085f4a43a878da8448615bc4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e36ff7a500ef444cbf5a168c5a48208b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 980.039305] env[61906]: DEBUG nova.network.neutron [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 980.043451] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-623cee29-dcb0-42fb-93b1-6517c22dda48 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.053606] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46384752-39d6-44a3-8f6c-31392a341fe5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.088111] env[61906]: DEBUG oslo_vmware.rw_handles [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ef10c7-d4bc-2624-0865-b110aac993bd/disk-0.vmdk. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 980.088922] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c914b84b-e3fd-448c-b3a5-35df8e5a161f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.091898] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daee2aff-c558-4ce0-aef3-30c214e64c80 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.102855] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a54347f0-0e38-4e8a-8960-248e724383e1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.106697] env[61906]: DEBUG oslo_vmware.rw_handles [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ef10c7-d4bc-2624-0865-b110aac993bd/disk-0.vmdk is in state: ready. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 980.106874] env[61906]: ERROR oslo_vmware.rw_handles [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ef10c7-d4bc-2624-0865-b110aac993bd/disk-0.vmdk due to incomplete transfer. [ 980.107114] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-46bcf67d-dad9-4891-a971-d5e4290acd34 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.118093] env[61906]: DEBUG nova.compute.provider_tree [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.120272] env[61906]: DEBUG oslo_vmware.rw_handles [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ef10c7-d4bc-2624-0865-b110aac993bd/disk-0.vmdk. {{(pid=61906) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 980.120468] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Uploaded image 40b9307b-1417-46aa-b262-7664cfccd96f to the Glance image server {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 980.125816] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Destroying the VM {{(pid=61906) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 980.125816] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-429dca62-fc17-4361-84b3-b14786d7d4b4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.128456] env[61906]: DEBUG oslo_vmware.api [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 980.128456] env[61906]: value = "task-1333497" [ 980.128456] env[61906]: _type = "Task" [ 980.128456] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.137492] env[61906]: DEBUG oslo_vmware.api [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333497, 'name': Destroy_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.193384] env[61906]: DEBUG nova.network.neutron [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Successfully created port: 73d0e617-fd5e-4fa2-81c5-710a3beac1a4 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 980.214691] env[61906]: DEBUG nova.network.neutron [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Updating instance_info_cache with network_info: [{"id": "e3ce81bf-b27e-47ea-9c48-a17e608a9cd6", "address": "fa:16:3e:03:ef:06", "network": {"id": "ddc08dc9-d08f-4fb9-9bee-1a030ed69f76", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1124067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b33dea0794a48f78b9f519cb269a8c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f1b07b1-e4e5-4842-9090-07fb2c3e124b", "external-id": "nsx-vlan-transportzone-646", "segmentation_id": 646, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3ce81bf-b2", "ovs_interfaceid": "e3ce81bf-b27e-47ea-9c48-a17e608a9cd6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.237042] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333496, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.309801] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52bc96c5-ca66-a38e-882f-3112efc372c4, 'name': SearchDatastore_Task, 'duration_secs': 0.495838} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.310602] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c256efb-0817-4d12-a770-00b15e5b8851 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.316486] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 980.316486] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52a4ee18-1a28-07c9-ebea-d732155d2970" [ 980.316486] env[61906]: _type = "Task" [ 980.316486] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.325617] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52a4ee18-1a28-07c9-ebea-d732155d2970, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.365953] env[61906]: DEBUG nova.compute.manager [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 980.624302] env[61906]: DEBUG nova.scheduler.client.report [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 980.638533] env[61906]: DEBUG oslo_vmware.api [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333497, 'name': Destroy_Task} progress is 33%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.717567] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Releasing lock "refresh_cache-a3e3ad00-6921-4072-8cb1-d80302883513" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.717912] env[61906]: DEBUG nova.compute.manager [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Instance network_info: |[{"id": "e3ce81bf-b27e-47ea-9c48-a17e608a9cd6", "address": "fa:16:3e:03:ef:06", "network": {"id": "ddc08dc9-d08f-4fb9-9bee-1a030ed69f76", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1124067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b33dea0794a48f78b9f519cb269a8c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f1b07b1-e4e5-4842-9090-07fb2c3e124b", "external-id": "nsx-vlan-transportzone-646", "segmentation_id": 646, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3ce81bf-b2", "ovs_interfaceid": "e3ce81bf-b27e-47ea-9c48-a17e608a9cd6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 980.718362] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:ef:06', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6f1b07b1-e4e5-4842-9090-07fb2c3e124b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e3ce81bf-b27e-47ea-9c48-a17e608a9cd6', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 980.726111] env[61906]: DEBUG oslo.service.loopingcall [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 980.726354] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 980.726590] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-077f9967-b2df-49fd-949f-beb2301aa1e6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.748875] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333496, 'name': ReconfigVM_Task, 'duration_secs': 1.872301} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.750065] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Reconfigured VM instance instance-00000057 to attach disk [datastore1] f49fb403-fbb4-4e26-8e75-c160d11dea05/f49fb403-fbb4-4e26-8e75-c160d11dea05.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 980.750666] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 980.750666] env[61906]: value = "task-1333498" [ 980.750666] env[61906]: _type = "Task" [ 980.750666] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.750866] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-154d5733-9aed-4c15-9302-2accac94afa2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.760022] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333498, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.760786] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 980.760786] env[61906]: value = "task-1333499" [ 980.760786] env[61906]: _type = "Task" [ 980.760786] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.771437] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333499, 'name': Rename_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.827591] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52a4ee18-1a28-07c9-ebea-d732155d2970, 'name': SearchDatastore_Task, 'duration_secs': 0.011029} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.828332] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.828592] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 458a1669-a62b-4313-874a-e49809d5c034/458a1669-a62b-4313-874a-e49809d5c034.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 980.828817] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-27507f69-8380-4116-911b-76b912f153b1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.835543] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 980.835543] env[61906]: value = "task-1333500" [ 980.835543] env[61906]: _type = "Task" [ 980.835543] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.843773] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333500, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.130197] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.775s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.134682] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.188s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.134682] env[61906]: DEBUG nova.objects.instance [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lazy-loading 'resources' on Instance uuid bddfdc7f-09d7-4887-b330-34a596ffa562 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 981.146286] env[61906]: DEBUG oslo_vmware.api [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333497, 'name': Destroy_Task, 'duration_secs': 0.932631} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.147917] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Destroyed the VM [ 981.147917] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Deleting Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 981.147917] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b9bafeec-79d7-48c0-9bcb-c6b45b53e528 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.157265] env[61906]: DEBUG oslo_vmware.api [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 981.157265] env[61906]: value = "task-1333501" [ 981.157265] env[61906]: _type = "Task" [ 981.157265] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.158415] env[61906]: INFO nova.scheduler.client.report [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Deleted allocations for instance 7142e219-6150-4c6f-9632-cbf489035431 [ 981.172434] env[61906]: DEBUG oslo_vmware.api [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333501, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.265079] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333498, 'name': CreateVM_Task, 'duration_secs': 0.492297} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.268539] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 981.269351] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.269529] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.269917] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 981.270679] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3b866d2-34a8-4d4c-8579-30f72a336451 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.276383] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333499, 'name': Rename_Task, 'duration_secs': 0.193683} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.277094] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 981.277421] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2b768b65-5b2f-43c9-ba08-573133cc7297 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.282468] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 981.282468] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5200712d-b2c5-d3dc-ad8e-dd230fb7a13c" [ 981.282468] env[61906]: _type = "Task" [ 981.282468] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.287284] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 981.287284] env[61906]: value = "task-1333502" [ 981.287284] env[61906]: _type = "Task" [ 981.287284] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.299688] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5200712d-b2c5-d3dc-ad8e-dd230fb7a13c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.303394] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333502, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.348923] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333500, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.376796] env[61906]: DEBUG nova.compute.manager [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 981.402174] env[61906]: DEBUG nova.virt.hardware [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 981.402519] env[61906]: DEBUG nova.virt.hardware [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 981.402729] env[61906]: DEBUG nova.virt.hardware [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 981.402938] env[61906]: DEBUG nova.virt.hardware [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 981.403115] env[61906]: DEBUG nova.virt.hardware [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 981.403276] env[61906]: DEBUG nova.virt.hardware [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 981.403491] env[61906]: DEBUG nova.virt.hardware [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 981.403661] env[61906]: DEBUG nova.virt.hardware [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 981.403846] env[61906]: DEBUG nova.virt.hardware [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 981.404014] env[61906]: DEBUG nova.virt.hardware [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 981.404201] env[61906]: DEBUG nova.virt.hardware [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 981.405164] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bed4708a-8902-4989-8562-ebfaed1f89b1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.413541] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfe66f61-649e-49ad-b741-f0780f9f91f9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.638507] env[61906]: DEBUG nova.compute.manager [req-240c81e3-a987-4e82-bb1d-079e9945e18c req-616247de-fa60-4ec3-8346-4088bbdfc5dd service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Received event network-vif-plugged-73d0e617-fd5e-4fa2-81c5-710a3beac1a4 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 981.638843] env[61906]: DEBUG oslo_concurrency.lockutils [req-240c81e3-a987-4e82-bb1d-079e9945e18c req-616247de-fa60-4ec3-8346-4088bbdfc5dd service nova] Acquiring lock "7264af54-cd10-4507-8ae3-5bb8a5cd127f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.639305] env[61906]: DEBUG oslo_concurrency.lockutils [req-240c81e3-a987-4e82-bb1d-079e9945e18c req-616247de-fa60-4ec3-8346-4088bbdfc5dd service nova] Lock "7264af54-cd10-4507-8ae3-5bb8a5cd127f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.639524] env[61906]: DEBUG oslo_concurrency.lockutils [req-240c81e3-a987-4e82-bb1d-079e9945e18c req-616247de-fa60-4ec3-8346-4088bbdfc5dd service nova] Lock "7264af54-cd10-4507-8ae3-5bb8a5cd127f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.639812] env[61906]: DEBUG nova.compute.manager [req-240c81e3-a987-4e82-bb1d-079e9945e18c req-616247de-fa60-4ec3-8346-4088bbdfc5dd service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] No waiting events found dispatching network-vif-plugged-73d0e617-fd5e-4fa2-81c5-710a3beac1a4 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 981.639894] env[61906]: WARNING nova.compute.manager [req-240c81e3-a987-4e82-bb1d-079e9945e18c req-616247de-fa60-4ec3-8346-4088bbdfc5dd service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Received unexpected event network-vif-plugged-73d0e617-fd5e-4fa2-81c5-710a3beac1a4 for instance with vm_state building and task_state spawning. [ 981.671447] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9eb99a1a-a126-4768-a8ba-199273a16f98 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "7142e219-6150-4c6f-9632-cbf489035431" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.191s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.678532] env[61906]: DEBUG oslo_vmware.api [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333501, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.754085] env[61906]: DEBUG nova.network.neutron [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Successfully updated port: 73d0e617-fd5e-4fa2-81c5-710a3beac1a4 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 981.795925] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5200712d-b2c5-d3dc-ad8e-dd230fb7a13c, 'name': SearchDatastore_Task, 'duration_secs': 0.064792} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.798301] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.798547] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 981.798782] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.798936] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.799134] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 981.804074] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-07e5911d-d678-4783-bca2-97f2a0c8c1d6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.809395] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333502, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.819615] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 981.819820] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 981.821627] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b126d71b-f03b-45f0-8a4c-de30b6365a34 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.826735] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 981.826735] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]529a13e4-9223-d8c6-2e15-fc7e008bf7bf" [ 981.826735] env[61906]: _type = "Task" [ 981.826735] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.834851] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]529a13e4-9223-d8c6-2e15-fc7e008bf7bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.845298] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333500, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.602975} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.845748] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 458a1669-a62b-4313-874a-e49809d5c034/458a1669-a62b-4313-874a-e49809d5c034.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 981.845836] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 981.847042] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-013eb3ea-4ea6-4abe-8248-5a0571c1ee06 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.849144] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-257d4ad6-c764-4e5c-b219-598fe2c205d3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.855908] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de93e3ce-3d49-4b17-adb5-c8bdd5ca3fc2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.860999] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 981.860999] env[61906]: value = "task-1333503" [ 981.860999] env[61906]: _type = "Task" [ 981.860999] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.864516] env[61906]: DEBUG oslo_concurrency.lockutils [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "37124c5c-2021-415b-925b-52e23a3d2973" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.864747] env[61906]: DEBUG oslo_concurrency.lockutils [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "37124c5c-2021-415b-925b-52e23a3d2973" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.897025] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd20e553-1bcc-497b-a66d-011c4b82418d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.902847] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333503, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.907811] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77bd7def-73ff-4b0a-b93e-b3d86b2f001d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.924096] env[61906]: DEBUG nova.compute.provider_tree [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 982.170964] env[61906]: DEBUG oslo_vmware.api [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333501, 'name': RemoveSnapshot_Task, 'duration_secs': 0.777112} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.171269] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Deleted Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 982.171514] env[61906]: INFO nova.compute.manager [None req-7efd9cae-d617-4c5a-9376-f131d61c7308 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Took 15.48 seconds to snapshot the instance on the hypervisor. [ 982.261105] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.261105] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.261254] env[61906]: DEBUG nova.network.neutron [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 982.299019] env[61906]: DEBUG oslo_vmware.api [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333502, 'name': PowerOnVM_Task, 'duration_secs': 0.84763} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.299019] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 982.299019] env[61906]: INFO nova.compute.manager [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Took 10.36 seconds to spawn the instance on the hypervisor. [ 982.299019] env[61906]: DEBUG nova.compute.manager [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 982.299753] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ebcc4fc-4d90-451e-a4db-49cecf02758d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.337753] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]529a13e4-9223-d8c6-2e15-fc7e008bf7bf, 'name': SearchDatastore_Task, 'duration_secs': 0.010159} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.338575] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8715e586-c8e6-4352-8a20-b686d905b07c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.344264] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 982.344264] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5265ed53-efb8-2c04-8d74-ca650197bed1" [ 982.344264] env[61906]: _type = "Task" [ 982.344264] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.354203] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5265ed53-efb8-2c04-8d74-ca650197bed1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.372839] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333503, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070325} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.373146] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 982.373915] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd1c4efb-ae1d-49f9-aa9d-024ab9fd8fb3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.396619] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 458a1669-a62b-4313-874a-e49809d5c034/458a1669-a62b-4313-874a-e49809d5c034.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 982.397083] env[61906]: DEBUG nova.compute.manager [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 982.399511] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b937bc36-e0a1-405f-a678-2ccbd30f6b79 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.420416] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 982.420416] env[61906]: value = "task-1333504" [ 982.420416] env[61906]: _type = "Task" [ 982.420416] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.429186] env[61906]: DEBUG nova.scheduler.client.report [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 982.432250] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333504, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.791638] env[61906]: DEBUG nova.network.neutron [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 982.820120] env[61906]: INFO nova.compute.manager [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Took 26.11 seconds to build instance. [ 982.858035] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5265ed53-efb8-2c04-8d74-ca650197bed1, 'name': SearchDatastore_Task, 'duration_secs': 0.050378} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.858794] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.858994] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] a3e3ad00-6921-4072-8cb1-d80302883513/a3e3ad00-6921-4072-8cb1-d80302883513.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 982.859503] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f52f2c21-3c19-4582-a229-cf6ca7bf9dfe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.867945] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 982.867945] env[61906]: value = "task-1333505" [ 982.867945] env[61906]: _type = "Task" [ 982.867945] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.876637] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333505, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.933538] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.799s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.935546] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333504, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.937167] env[61906]: DEBUG oslo_concurrency.lockutils [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.937167] env[61906]: DEBUG oslo_concurrency.lockutils [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.940832] env[61906]: INFO nova.compute.claims [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 982.963322] env[61906]: INFO nova.scheduler.client.report [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Deleted allocations for instance bddfdc7f-09d7-4887-b330-34a596ffa562 [ 982.998016] env[61906]: DEBUG nova.network.neutron [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Updating instance_info_cache with network_info: [{"id": "73d0e617-fd5e-4fa2-81c5-710a3beac1a4", "address": "fa:16:3e:66:bf:d9", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73d0e617-fd", "ovs_interfaceid": "73d0e617-fd5e-4fa2-81c5-710a3beac1a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.320798] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "f49fb403-fbb4-4e26-8e75-c160d11dea05" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.323279] env[61906]: DEBUG oslo_concurrency.lockutils [None req-217a7949-cdd6-4b30-a0bf-40a892e1b338 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "f49fb403-fbb4-4e26-8e75-c160d11dea05" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.621s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.325628] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "f49fb403-fbb4-4e26-8e75-c160d11dea05" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.003s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.325628] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "f49fb403-fbb4-4e26-8e75-c160d11dea05-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.325628] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "f49fb403-fbb4-4e26-8e75-c160d11dea05-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.325628] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "f49fb403-fbb4-4e26-8e75-c160d11dea05-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.327156] env[61906]: INFO nova.compute.manager [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Terminating instance [ 983.328357] env[61906]: DEBUG nova.compute.manager [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 983.330034] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 983.330034] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5187190-cda8-4f6f-87d1-2311ae6982c3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.337850] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 983.338123] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-87e14420-04e4-4ed0-95b0-f872ed5e9241 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.346200] env[61906]: DEBUG oslo_vmware.api [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 983.346200] env[61906]: value = "task-1333506" [ 983.346200] env[61906]: _type = "Task" [ 983.346200] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.360132] env[61906]: DEBUG oslo_vmware.api [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333506, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.377996] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333505, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.434444] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333504, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.476959] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6463c81-37a4-474e-959d-b1d1315d23ec tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "bddfdc7f-09d7-4887-b330-34a596ffa562" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.931s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.500100] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Releasing lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.500456] env[61906]: DEBUG nova.compute.manager [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Instance network_info: |[{"id": "73d0e617-fd5e-4fa2-81c5-710a3beac1a4", "address": "fa:16:3e:66:bf:d9", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73d0e617-fd", "ovs_interfaceid": "73d0e617-fd5e-4fa2-81c5-710a3beac1a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 983.500933] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:66:bf:d9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35ac9709-fd8b-4630-897a-68ed629d1b11', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '73d0e617-fd5e-4fa2-81c5-710a3beac1a4', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 983.509542] env[61906]: DEBUG oslo.service.loopingcall [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 983.510212] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 983.510458] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0031604f-0714-4107-98a0-11ac9f64bb44 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.533544] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 983.533544] env[61906]: value = "task-1333507" [ 983.533544] env[61906]: _type = "Task" [ 983.533544] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.544045] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333507, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.688395] env[61906]: DEBUG nova.compute.manager [req-091104fc-6da2-4a7f-8df9-ab238d3b6b03 req-f429a7a0-c446-4b2f-b8fe-10b8d3f64d33 service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Received event network-changed-73d0e617-fd5e-4fa2-81c5-710a3beac1a4 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 983.689028] env[61906]: DEBUG nova.compute.manager [req-091104fc-6da2-4a7f-8df9-ab238d3b6b03 req-f429a7a0-c446-4b2f-b8fe-10b8d3f64d33 service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Refreshing instance network info cache due to event network-changed-73d0e617-fd5e-4fa2-81c5-710a3beac1a4. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 983.689417] env[61906]: DEBUG oslo_concurrency.lockutils [req-091104fc-6da2-4a7f-8df9-ab238d3b6b03 req-f429a7a0-c446-4b2f-b8fe-10b8d3f64d33 service nova] Acquiring lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.689681] env[61906]: DEBUG oslo_concurrency.lockutils [req-091104fc-6da2-4a7f-8df9-ab238d3b6b03 req-f429a7a0-c446-4b2f-b8fe-10b8d3f64d33 service nova] Acquired lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.689968] env[61906]: DEBUG nova.network.neutron [req-091104fc-6da2-4a7f-8df9-ab238d3b6b03 req-f429a7a0-c446-4b2f-b8fe-10b8d3f64d33 service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Refreshing network info cache for port 73d0e617-fd5e-4fa2-81c5-710a3beac1a4 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 983.856691] env[61906]: DEBUG oslo_vmware.api [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333506, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.878557] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333505, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.930459} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.878783] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] a3e3ad00-6921-4072-8cb1-d80302883513/a3e3ad00-6921-4072-8cb1-d80302883513.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 983.878993] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 983.879284] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3095233c-a046-4021-82fd-0c9c3a493b53 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.887880] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 983.887880] env[61906]: value = "task-1333508" [ 983.887880] env[61906]: _type = "Task" [ 983.887880] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.898035] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333508, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.934813] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333504, 'name': ReconfigVM_Task, 'duration_secs': 1.299929} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.935083] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 458a1669-a62b-4313-874a-e49809d5c034/458a1669-a62b-4313-874a-e49809d5c034.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 983.935772] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-81bc6655-00c0-4f9a-9fd7-1d19531f6070 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.949542] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 983.949542] env[61906]: value = "task-1333509" [ 983.949542] env[61906]: _type = "Task" [ 983.949542] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.960107] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333509, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.046960] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333507, 'name': CreateVM_Task, 'duration_secs': 0.480488} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.050433] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 984.053058] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.053411] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.053874] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 984.054294] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1531809d-e930-4232-af74-e3695a4a843c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.062767] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 984.062767] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52782181-2c64-096a-29f3-317f32ca6ea9" [ 984.062767] env[61906]: _type = "Task" [ 984.062767] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.076820] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52782181-2c64-096a-29f3-317f32ca6ea9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.172209] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bedc4c2b-c6e1-4e33-b5ea-053a82546db2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.180959] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bdf093c-1818-460d-ba1a-2545d8509a39 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.217096] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b23a0d-3a4f-430a-945f-642e7c681860 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.225593] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae17a454-21fa-4823-9aef-931b39b0ff1c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.242400] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "34d22e43-882f-45c2-8259-e77424ce0fb6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.242816] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "34d22e43-882f-45c2-8259-e77424ce0fb6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.244250] env[61906]: DEBUG nova.compute.provider_tree [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.266754] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "cec87a1e-4176-45bb-a0c7-a594cb399170" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.266985] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "cec87a1e-4176-45bb-a0c7-a594cb399170" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.356478] env[61906]: DEBUG oslo_vmware.api [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333506, 'name': PowerOffVM_Task, 'duration_secs': 0.516245} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.356713] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 984.356889] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 984.357163] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0aad2418-3764-4c22-92d6-f01801251aed {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.400672] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333508, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065194} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.400957] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 984.401840] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-419aa6c3-e33d-4fc0-8117-427e0ec4a1b0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.425733] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] a3e3ad00-6921-4072-8cb1-d80302883513/a3e3ad00-6921-4072-8cb1-d80302883513.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 984.426097] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e9218e0-94db-438c-aab7-abe7876b6eda {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.441947] env[61906]: DEBUG nova.network.neutron [req-091104fc-6da2-4a7f-8df9-ab238d3b6b03 req-f429a7a0-c446-4b2f-b8fe-10b8d3f64d33 service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Updated VIF entry in instance network info cache for port 73d0e617-fd5e-4fa2-81c5-710a3beac1a4. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 984.442286] env[61906]: DEBUG nova.network.neutron [req-091104fc-6da2-4a7f-8df9-ab238d3b6b03 req-f429a7a0-c446-4b2f-b8fe-10b8d3f64d33 service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Updating instance_info_cache with network_info: [{"id": "73d0e617-fd5e-4fa2-81c5-710a3beac1a4", "address": "fa:16:3e:66:bf:d9", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73d0e617-fd", "ovs_interfaceid": "73d0e617-fd5e-4fa2-81c5-710a3beac1a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.444543] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 984.444750] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 984.444940] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Deleting the datastore file [datastore1] f49fb403-fbb4-4e26-8e75-c160d11dea05 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 984.445751] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a351ac47-9e48-416e-a4ce-7f5f279d297a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.451669] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 984.451669] env[61906]: value = "task-1333511" [ 984.451669] env[61906]: _type = "Task" [ 984.451669] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.456712] env[61906]: DEBUG oslo_vmware.api [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 984.456712] env[61906]: value = "task-1333512" [ 984.456712] env[61906]: _type = "Task" [ 984.456712] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.466649] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333509, 'name': Rename_Task, 'duration_secs': 0.282112} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.467265] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 984.467521] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5ffad9e2-8c1b-4df4-975b-1f4753219776 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.475462] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333511, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.476016] env[61906]: DEBUG oslo_vmware.api [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333512, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.477224] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 984.477224] env[61906]: value = "task-1333513" [ 984.477224] env[61906]: _type = "Task" [ 984.477224] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.485127] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333513, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.574339] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52782181-2c64-096a-29f3-317f32ca6ea9, 'name': SearchDatastore_Task, 'duration_secs': 0.010751} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.574673] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.574947] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 984.575216] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.575371] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.575552] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 984.575821] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6f324e94-3daa-4bb4-bd65-fd54b9bfe211 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.587496] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 984.587678] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 984.588407] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd5d6c10-6984-4ab4-8de8-dfde9d6654a4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.594445] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 984.594445] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5275e28d-1b42-dec8-3c45-b5be8b1837ef" [ 984.594445] env[61906]: _type = "Task" [ 984.594445] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.602206] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5275e28d-1b42-dec8-3c45-b5be8b1837ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.746866] env[61906]: DEBUG nova.compute.manager [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 984.750153] env[61906]: DEBUG nova.scheduler.client.report [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 984.768792] env[61906]: DEBUG nova.compute.manager [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 984.946526] env[61906]: DEBUG oslo_concurrency.lockutils [req-091104fc-6da2-4a7f-8df9-ab238d3b6b03 req-f429a7a0-c446-4b2f-b8fe-10b8d3f64d33 service nova] Releasing lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.969254] env[61906]: DEBUG oslo_vmware.api [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333512, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.465257} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.972304] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 984.972506] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 984.972708] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 984.972944] env[61906]: INFO nova.compute.manager [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Took 1.64 seconds to destroy the instance on the hypervisor. [ 984.973212] env[61906]: DEBUG oslo.service.loopingcall [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 984.973374] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333511, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.973589] env[61906]: DEBUG nova.compute.manager [-] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 984.973700] env[61906]: DEBUG nova.network.neutron [-] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 984.986237] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333513, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.107744] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5275e28d-1b42-dec8-3c45-b5be8b1837ef, 'name': SearchDatastore_Task, 'duration_secs': 0.019834} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.108575] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69b9f834-1423-49b8-ba20-c1bec6b5d3d0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.115417] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 985.115417] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5229b1cd-27cf-6ee4-72be-d266749c5c07" [ 985.115417] env[61906]: _type = "Task" [ 985.115417] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.125780] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5229b1cd-27cf-6ee4-72be-d266749c5c07, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.257181] env[61906]: DEBUG oslo_concurrency.lockutils [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.320s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.257667] env[61906]: DEBUG nova.compute.manager [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 985.282374] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.282718] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.285027] env[61906]: INFO nova.compute.claims [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 985.292217] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.467960] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333511, 'name': ReconfigVM_Task, 'duration_secs': 0.862953} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.468300] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Reconfigured VM instance instance-00000059 to attach disk [datastore1] a3e3ad00-6921-4072-8cb1-d80302883513/a3e3ad00-6921-4072-8cb1-d80302883513.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 985.468958] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f2d4de92-347a-4362-86be-f41aab574850 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.477484] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 985.477484] env[61906]: value = "task-1333514" [ 985.477484] env[61906]: _type = "Task" [ 985.477484] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.492990] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333514, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.493300] env[61906]: DEBUG oslo_vmware.api [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333513, 'name': PowerOnVM_Task, 'duration_secs': 0.708132} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.493537] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 985.493794] env[61906]: INFO nova.compute.manager [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Took 10.44 seconds to spawn the instance on the hypervisor. [ 985.493963] env[61906]: DEBUG nova.compute.manager [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 985.494749] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6ebbeac-c53b-4cda-899b-eed6c28f7fb0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.626617] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5229b1cd-27cf-6ee4-72be-d266749c5c07, 'name': SearchDatastore_Task, 'duration_secs': 0.010176} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.626884] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.627169] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 7264af54-cd10-4507-8ae3-5bb8a5cd127f/7264af54-cd10-4507-8ae3-5bb8a5cd127f.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 985.627429] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7170745a-88eb-4e61-9e0e-c7e71fa100d8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.633715] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 985.633715] env[61906]: value = "task-1333515" [ 985.633715] env[61906]: _type = "Task" [ 985.633715] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.640934] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333515, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.716222] env[61906]: DEBUG nova.compute.manager [req-4ffb4651-64ce-486e-91d8-1b8266d8b09e req-b28600a8-5739-4f5d-9b48-acbe5859bccf service nova] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Received event network-vif-deleted-a8e2f461-ff1a-4839-a351-a9a1925b7be2 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 985.716436] env[61906]: INFO nova.compute.manager [req-4ffb4651-64ce-486e-91d8-1b8266d8b09e req-b28600a8-5739-4f5d-9b48-acbe5859bccf service nova] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Neutron deleted interface a8e2f461-ff1a-4839-a351-a9a1925b7be2; detaching it from the instance and deleting it from the info cache [ 985.716615] env[61906]: DEBUG nova.network.neutron [req-4ffb4651-64ce-486e-91d8-1b8266d8b09e req-b28600a8-5739-4f5d-9b48-acbe5859bccf service nova] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.764369] env[61906]: DEBUG nova.compute.utils [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 985.765503] env[61906]: DEBUG nova.compute.manager [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 985.765706] env[61906]: DEBUG nova.network.neutron [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 985.769210] env[61906]: DEBUG nova.network.neutron [-] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.811009] env[61906]: DEBUG nova.policy [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc6b374325bc4ef9882e5f84ddd62fd1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f2d2198383d74dac90aa9727a62176df', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 985.988073] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333514, 'name': Rename_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.013544] env[61906]: INFO nova.compute.manager [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Took 20.98 seconds to build instance. [ 986.094437] env[61906]: DEBUG nova.network.neutron [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Successfully created port: 077e6829-d958-4859-903a-8cad7602c935 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 986.145491] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333515, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.219114] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c0fd29b4-c73e-49b0-8cb0-1c326e9582e0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.230366] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2d75a8f-f2b0-4095-9942-7fde947a9514 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.263550] env[61906]: DEBUG nova.compute.manager [req-4ffb4651-64ce-486e-91d8-1b8266d8b09e req-b28600a8-5739-4f5d-9b48-acbe5859bccf service nova] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Detach interface failed, port_id=a8e2f461-ff1a-4839-a351-a9a1925b7be2, reason: Instance f49fb403-fbb4-4e26-8e75-c160d11dea05 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 986.268383] env[61906]: DEBUG nova.compute.manager [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 986.274172] env[61906]: INFO nova.compute.manager [-] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Took 1.30 seconds to deallocate network for instance. [ 986.488792] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333514, 'name': Rename_Task, 'duration_secs': 0.926879} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.490027] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 986.490828] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ddfc0d1-99b6-434f-9b07-5ea82824ab8a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.493459] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8050ea77-d584-4b00-9020-952ac1815490 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.502395] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f6b5d51-b09f-4a47-9b9f-941c29f7926e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.505831] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 986.505831] env[61906]: value = "task-1333516" [ 986.505831] env[61906]: _type = "Task" [ 986.505831] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.540201] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc4bda09-d170-4eed-b482-36c891da6c79 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "458a1669-a62b-4313-874a-e49809d5c034" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.514s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.541957] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba66814e-19c2-41bf-9060-8e20432a1fc9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.548579] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333516, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.555416] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e622648f-6c06-44a7-b084-b4852288e1b8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.575673] env[61906]: DEBUG nova.compute.provider_tree [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.645765] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333515, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.780316] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.018359] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333516, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.078386] env[61906]: DEBUG nova.scheduler.client.report [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 987.146080] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333515, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.166116} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.146348] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 7264af54-cd10-4507-8ae3-5bb8a5cd127f/7264af54-cd10-4507-8ae3-5bb8a5cd127f.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 987.146569] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 987.146828] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-07721b49-4d06-49cc-8a0a-5d5224e24609 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.153840] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 987.153840] env[61906]: value = "task-1333517" [ 987.153840] env[61906]: _type = "Task" [ 987.153840] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.161639] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333517, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.280363] env[61906]: DEBUG nova.compute.manager [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 987.305644] env[61906]: DEBUG nova.virt.hardware [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='c62fc15a70adbb4dd677213c36562064',container_format='bare',created_at=2024-10-22T06:26:57Z,direct_url=,disk_format='vmdk',id=40b9307b-1417-46aa-b262-7664cfccd96f,min_disk=1,min_ram=0,name='tempest-test-snap-829224891',owner='f2d2198383d74dac90aa9727a62176df',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-22T06:27:12Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 987.305925] env[61906]: DEBUG nova.virt.hardware [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 987.306106] env[61906]: DEBUG nova.virt.hardware [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 987.306300] env[61906]: DEBUG nova.virt.hardware [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 987.306454] env[61906]: DEBUG nova.virt.hardware [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 987.306606] env[61906]: DEBUG nova.virt.hardware [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 987.306841] env[61906]: DEBUG nova.virt.hardware [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 987.307104] env[61906]: DEBUG nova.virt.hardware [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 987.307301] env[61906]: DEBUG nova.virt.hardware [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 987.307475] env[61906]: DEBUG nova.virt.hardware [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 987.307656] env[61906]: DEBUG nova.virt.hardware [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 987.308526] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6ab7a9b-7ff3-4639-8cd2-e7703f230b11 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.317027] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8146afd9-11ad-46de-a8da-ef3142b9f4ee {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.514663] env[61906]: DEBUG nova.compute.manager [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Stashing vm_state: active {{(pid=61906) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 987.520655] env[61906]: DEBUG oslo_vmware.api [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333516, 'name': PowerOnVM_Task, 'duration_secs': 0.566939} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.520985] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 987.521216] env[61906]: INFO nova.compute.manager [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Took 10.17 seconds to spawn the instance on the hypervisor. [ 987.521402] env[61906]: DEBUG nova.compute.manager [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 987.522209] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df4065f6-d005-485b-a6f6-70eed6ddfb97 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.583200] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.300s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.583704] env[61906]: DEBUG nova.compute.manager [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 987.586070] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.294s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.587706] env[61906]: INFO nova.compute.claims [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 987.664592] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333517, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062328} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.664877] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 987.665681] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f6b8bc5-6612-44ed-b921-65753de0d242 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.687822] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] 7264af54-cd10-4507-8ae3-5bb8a5cd127f/7264af54-cd10-4507-8ae3-5bb8a5cd127f.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 987.688165] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f8049d0-313d-4884-9085-96e977c720a8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.710183] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 987.710183] env[61906]: value = "task-1333518" [ 987.710183] env[61906]: _type = "Task" [ 987.710183] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.718582] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333518, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.748471] env[61906]: DEBUG nova.compute.manager [req-69df2806-65f0-4605-b325-9fb2ea2e3b78 req-971578f8-2bef-4ea8-b935-6da70c9d7e1e service nova] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Received event network-vif-plugged-077e6829-d958-4859-903a-8cad7602c935 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 987.748916] env[61906]: DEBUG oslo_concurrency.lockutils [req-69df2806-65f0-4605-b325-9fb2ea2e3b78 req-971578f8-2bef-4ea8-b935-6da70c9d7e1e service nova] Acquiring lock "37124c5c-2021-415b-925b-52e23a3d2973-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.749664] env[61906]: DEBUG oslo_concurrency.lockutils [req-69df2806-65f0-4605-b325-9fb2ea2e3b78 req-971578f8-2bef-4ea8-b935-6da70c9d7e1e service nova] Lock "37124c5c-2021-415b-925b-52e23a3d2973-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.749971] env[61906]: DEBUG oslo_concurrency.lockutils [req-69df2806-65f0-4605-b325-9fb2ea2e3b78 req-971578f8-2bef-4ea8-b935-6da70c9d7e1e service nova] Lock "37124c5c-2021-415b-925b-52e23a3d2973-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.750269] env[61906]: DEBUG nova.compute.manager [req-69df2806-65f0-4605-b325-9fb2ea2e3b78 req-971578f8-2bef-4ea8-b935-6da70c9d7e1e service nova] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] No waiting events found dispatching network-vif-plugged-077e6829-d958-4859-903a-8cad7602c935 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 987.750478] env[61906]: WARNING nova.compute.manager [req-69df2806-65f0-4605-b325-9fb2ea2e3b78 req-971578f8-2bef-4ea8-b935-6da70c9d7e1e service nova] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Received unexpected event network-vif-plugged-077e6829-d958-4859-903a-8cad7602c935 for instance with vm_state building and task_state spawning. [ 987.791094] env[61906]: DEBUG nova.network.neutron [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Successfully updated port: 077e6829-d958-4859-903a-8cad7602c935 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 988.040474] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.041951] env[61906]: INFO nova.compute.manager [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Took 19.81 seconds to build instance. [ 988.090950] env[61906]: DEBUG nova.compute.utils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 988.094392] env[61906]: DEBUG nova.compute.manager [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 988.094533] env[61906]: DEBUG nova.network.neutron [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 988.133431] env[61906]: DEBUG nova.policy [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '23ff59c262484422a8badcbca5a103f9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '76ca0b5a4fd24ede8eac4851fc365848', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 988.221401] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333518, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.293521] env[61906]: DEBUG oslo_concurrency.lockutils [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "refresh_cache-37124c5c-2021-415b-925b-52e23a3d2973" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.293706] env[61906]: DEBUG oslo_concurrency.lockutils [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquired lock "refresh_cache-37124c5c-2021-415b-925b-52e23a3d2973" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.293925] env[61906]: DEBUG nova.network.neutron [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 988.374700] env[61906]: DEBUG nova.network.neutron [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Successfully created port: 5cee4cef-6d3a-4191-836c-a40217b08b3c {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 988.546241] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d9f157e2-69fb-4ef9-ab62-e383bf64798c tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "a3e3ad00-6921-4072-8cb1-d80302883513" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.320s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.595641] env[61906]: DEBUG nova.compute.manager [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 988.721817] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333518, 'name': ReconfigVM_Task, 'duration_secs': 0.771549} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.724596] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Reconfigured VM instance instance-0000005a to attach disk [datastore2] 7264af54-cd10-4507-8ae3-5bb8a5cd127f/7264af54-cd10-4507-8ae3-5bb8a5cd127f.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 988.725537] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-edbb4223-da3e-427f-9f0a-1d0957f70e2a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.732591] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 988.732591] env[61906]: value = "task-1333519" [ 988.732591] env[61906]: _type = "Task" [ 988.732591] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.743669] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333519, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.799417] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a31a0e42-28e8-49e3-acd0-2c859dffbf6d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.809356] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afb2e32e-3e96-430e-a1ac-6f1f3525f053 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.841498] env[61906]: DEBUG nova.network.neutron [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 988.843937] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290e95d9-b969-406e-b845-055d1e06035e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.852145] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bc0b62a-1232-47f8-aef7-04d5f3492e75 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.866392] env[61906]: DEBUG nova.compute.provider_tree [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 989.062226] env[61906]: DEBUG nova.network.neutron [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Updating instance_info_cache with network_info: [{"id": "077e6829-d958-4859-903a-8cad7602c935", "address": "fa:16:3e:b7:1e:9e", "network": {"id": "d123fdf4-6bb4-400b-a0bb-9d477672af2c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2099881656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d2198383d74dac90aa9727a62176df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap077e6829-d9", "ovs_interfaceid": "077e6829-d958-4859-903a-8cad7602c935", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.244793] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333519, 'name': Rename_Task, 'duration_secs': 0.335485} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.245022] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 989.245280] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ebd19b76-006c-4254-95eb-7425ff197981 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.252227] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 989.252227] env[61906]: value = "task-1333520" [ 989.252227] env[61906]: _type = "Task" [ 989.252227] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.259957] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333520, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.390695] env[61906]: ERROR nova.scheduler.client.report [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [req-7804ebab-3292-464f-ba88-6f222303c0b5] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID a5b57df7-a16e-44f9-9b2d-23c518860263. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-7804ebab-3292-464f-ba88-6f222303c0b5"}]} [ 989.407153] env[61906]: DEBUG nova.scheduler.client.report [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Refreshing inventories for resource provider a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 989.421662] env[61906]: DEBUG nova.scheduler.client.report [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Updating ProviderTree inventory for provider a5b57df7-a16e-44f9-9b2d-23c518860263 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 989.421905] env[61906]: DEBUG nova.compute.provider_tree [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 989.433795] env[61906]: DEBUG nova.scheduler.client.report [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Refreshing aggregate associations for resource provider a5b57df7-a16e-44f9-9b2d-23c518860263, aggregates: None {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 989.450045] env[61906]: DEBUG nova.scheduler.client.report [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Refreshing trait associations for resource provider a5b57df7-a16e-44f9-9b2d-23c518860263, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64 {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 989.565280] env[61906]: DEBUG oslo_concurrency.lockutils [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Releasing lock "refresh_cache-37124c5c-2021-415b-925b-52e23a3d2973" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.565617] env[61906]: DEBUG nova.compute.manager [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Instance network_info: |[{"id": "077e6829-d958-4859-903a-8cad7602c935", "address": "fa:16:3e:b7:1e:9e", "network": {"id": "d123fdf4-6bb4-400b-a0bb-9d477672af2c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2099881656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d2198383d74dac90aa9727a62176df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap077e6829-d9", "ovs_interfaceid": "077e6829-d958-4859-903a-8cad7602c935", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 989.566125] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b7:1e:9e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cba18f15-a919-422e-a423-1e705e233389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '077e6829-d958-4859-903a-8cad7602c935', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 989.574938] env[61906]: DEBUG oslo.service.loopingcall [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 989.577675] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 989.578112] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-839a1ab4-6425-4b12-bfdf-0d302fdb7c94 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.601658] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 989.601658] env[61906]: value = "task-1333521" [ 989.601658] env[61906]: _type = "Task" [ 989.601658] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.608029] env[61906]: DEBUG nova.compute.manager [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 989.614951] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333521, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.636792] env[61906]: DEBUG nova.virt.hardware [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 989.636792] env[61906]: DEBUG nova.virt.hardware [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 989.636792] env[61906]: DEBUG nova.virt.hardware [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 989.637111] env[61906]: DEBUG nova.virt.hardware [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 989.637267] env[61906]: DEBUG nova.virt.hardware [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 989.637547] env[61906]: DEBUG nova.virt.hardware [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 989.637721] env[61906]: DEBUG nova.virt.hardware [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 989.637927] env[61906]: DEBUG nova.virt.hardware [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 989.638153] env[61906]: DEBUG nova.virt.hardware [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 989.638410] env[61906]: DEBUG nova.virt.hardware [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 989.638868] env[61906]: DEBUG nova.virt.hardware [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 989.639734] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6f97da8-f98e-4772-895b-c2524a19a2ca {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.652389] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0395c47c-605c-4e9f-a4cd-a4d4aaaef24b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.658372] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e106e9-4262-4c82-a23b-5ead732d52db {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.675099] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19a0f9d3-eda1-479b-8a39-724dcff8a495 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.707951] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ed71081-36dc-4e51-a276-13daefbd36a5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.721978] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d06b822-3569-4045-b212-6c57ad1d6db1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.742828] env[61906]: DEBUG nova.compute.provider_tree [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 989.763952] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333520, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.778846] env[61906]: DEBUG nova.compute.manager [req-cfbc63c0-fef9-4100-8e53-d34ac7207097 req-f4e6bb24-264c-4a1e-bdf7-900d03fc74d0 service nova] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Received event network-changed-077e6829-d958-4859-903a-8cad7602c935 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.779076] env[61906]: DEBUG nova.compute.manager [req-cfbc63c0-fef9-4100-8e53-d34ac7207097 req-f4e6bb24-264c-4a1e-bdf7-900d03fc74d0 service nova] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Refreshing instance network info cache due to event network-changed-077e6829-d958-4859-903a-8cad7602c935. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 989.779329] env[61906]: DEBUG oslo_concurrency.lockutils [req-cfbc63c0-fef9-4100-8e53-d34ac7207097 req-f4e6bb24-264c-4a1e-bdf7-900d03fc74d0 service nova] Acquiring lock "refresh_cache-37124c5c-2021-415b-925b-52e23a3d2973" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.779617] env[61906]: DEBUG oslo_concurrency.lockutils [req-cfbc63c0-fef9-4100-8e53-d34ac7207097 req-f4e6bb24-264c-4a1e-bdf7-900d03fc74d0 service nova] Acquired lock "refresh_cache-37124c5c-2021-415b-925b-52e23a3d2973" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.779817] env[61906]: DEBUG nova.network.neutron [req-cfbc63c0-fef9-4100-8e53-d34ac7207097 req-f4e6bb24-264c-4a1e-bdf7-900d03fc74d0 service nova] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Refreshing network info cache for port 077e6829-d958-4859-903a-8cad7602c935 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 990.061188] env[61906]: DEBUG nova.network.neutron [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Successfully updated port: 5cee4cef-6d3a-4191-836c-a40217b08b3c {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 990.111676] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333521, 'name': CreateVM_Task, 'duration_secs': 0.382338} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.112046] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 990.112791] env[61906]: DEBUG oslo_concurrency.lockutils [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/40b9307b-1417-46aa-b262-7664cfccd96f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.112791] env[61906]: DEBUG oslo_concurrency.lockutils [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquired lock "[datastore2] devstack-image-cache_base/40b9307b-1417-46aa-b262-7664cfccd96f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.113172] env[61906]: DEBUG oslo_concurrency.lockutils [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/40b9307b-1417-46aa-b262-7664cfccd96f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 990.113441] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13e050a6-39b1-40d0-a041-b29bfa32cafc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.118456] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 990.118456] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]523cc6ae-8664-6f8f-9791-c833c115ba5d" [ 990.118456] env[61906]: _type = "Task" [ 990.118456] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.126979] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]523cc6ae-8664-6f8f-9791-c833c115ba5d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.220210] env[61906]: DEBUG nova.compute.manager [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Stashing vm_state: active {{(pid=61906) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 990.262641] env[61906]: DEBUG oslo_vmware.api [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333520, 'name': PowerOnVM_Task, 'duration_secs': 0.636959} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.262952] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 990.263192] env[61906]: INFO nova.compute.manager [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Took 8.89 seconds to spawn the instance on the hypervisor. [ 990.263381] env[61906]: DEBUG nova.compute.manager [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 990.264156] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0e79c8d-9130-4159-8a89-8491baf18afc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.276050] env[61906]: DEBUG nova.scheduler.client.report [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Updated inventory for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with generation 112 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 990.276312] env[61906]: DEBUG nova.compute.provider_tree [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Updating resource provider a5b57df7-a16e-44f9-9b2d-23c518860263 generation from 112 to 113 during operation: update_inventory {{(pid=61906) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 990.276500] env[61906]: DEBUG nova.compute.provider_tree [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 990.475028] env[61906]: DEBUG nova.network.neutron [req-cfbc63c0-fef9-4100-8e53-d34ac7207097 req-f4e6bb24-264c-4a1e-bdf7-900d03fc74d0 service nova] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Updated VIF entry in instance network info cache for port 077e6829-d958-4859-903a-8cad7602c935. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 990.475473] env[61906]: DEBUG nova.network.neutron [req-cfbc63c0-fef9-4100-8e53-d34ac7207097 req-f4e6bb24-264c-4a1e-bdf7-900d03fc74d0 service nova] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Updating instance_info_cache with network_info: [{"id": "077e6829-d958-4859-903a-8cad7602c935", "address": "fa:16:3e:b7:1e:9e", "network": {"id": "d123fdf4-6bb4-400b-a0bb-9d477672af2c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2099881656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d2198383d74dac90aa9727a62176df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap077e6829-d9", "ovs_interfaceid": "077e6829-d958-4859-903a-8cad7602c935", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.563625] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "refresh_cache-34d22e43-882f-45c2-8259-e77424ce0fb6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.563808] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquired lock "refresh_cache-34d22e43-882f-45c2-8259-e77424ce0fb6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.563974] env[61906]: DEBUG nova.network.neutron [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 990.630106] env[61906]: DEBUG oslo_concurrency.lockutils [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Releasing lock "[datastore2] devstack-image-cache_base/40b9307b-1417-46aa-b262-7664cfccd96f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.630476] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Processing image 40b9307b-1417-46aa-b262-7664cfccd96f {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 990.630767] env[61906]: DEBUG oslo_concurrency.lockutils [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/40b9307b-1417-46aa-b262-7664cfccd96f/40b9307b-1417-46aa-b262-7664cfccd96f.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.630977] env[61906]: DEBUG oslo_concurrency.lockutils [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquired lock "[datastore2] devstack-image-cache_base/40b9307b-1417-46aa-b262-7664cfccd96f/40b9307b-1417-46aa-b262-7664cfccd96f.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.631271] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 990.631578] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5c7af4fb-3f1d-4604-9392-3063f3fb9e31 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.655880] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 990.656119] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 990.656868] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32bdcf49-393a-413c-b714-42a0173a555a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.662315] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 990.662315] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]523eeb0c-d78a-1a70-cba6-4f1486c95e58" [ 990.662315] env[61906]: _type = "Task" [ 990.662315] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.671829] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]523eeb0c-d78a-1a70-cba6-4f1486c95e58, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.740558] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.781914] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.195s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.782114] env[61906]: DEBUG nova.compute.manager [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 990.784897] env[61906]: INFO nova.compute.manager [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Took 19.23 seconds to build instance. [ 990.786405] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.007s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.787177] env[61906]: DEBUG nova.objects.instance [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lazy-loading 'resources' on Instance uuid f49fb403-fbb4-4e26-8e75-c160d11dea05 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 990.978351] env[61906]: DEBUG oslo_concurrency.lockutils [req-cfbc63c0-fef9-4100-8e53-d34ac7207097 req-f4e6bb24-264c-4a1e-bdf7-900d03fc74d0 service nova] Releasing lock "refresh_cache-37124c5c-2021-415b-925b-52e23a3d2973" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.118859] env[61906]: DEBUG nova.network.neutron [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 991.173071] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Preparing fetch location {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 991.173353] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Fetch image to [datastore2] OSTACK_IMG_4f594878-e59c-4af4-a30e-045edbff7e1f/OSTACK_IMG_4f594878-e59c-4af4-a30e-045edbff7e1f.vmdk {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 991.173542] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Downloading stream optimized image 40b9307b-1417-46aa-b262-7664cfccd96f to [datastore2] OSTACK_IMG_4f594878-e59c-4af4-a30e-045edbff7e1f/OSTACK_IMG_4f594878-e59c-4af4-a30e-045edbff7e1f.vmdk on the data store datastore2 as vApp {{(pid=61906) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 991.173921] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Downloading image file data 40b9307b-1417-46aa-b262-7664cfccd96f to the ESX as VM named 'OSTACK_IMG_4f594878-e59c-4af4-a30e-045edbff7e1f' {{(pid=61906) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 991.251383] env[61906]: DEBUG oslo_vmware.rw_handles [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 991.251383] env[61906]: value = "resgroup-9" [ 991.251383] env[61906]: _type = "ResourcePool" [ 991.251383] env[61906]: }. {{(pid=61906) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 991.251716] env[61906]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-75cb08e6-3173-4501-99f5-789a8d08d91a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.273331] env[61906]: DEBUG oslo_vmware.rw_handles [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lease: (returnval){ [ 991.273331] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52783e34-a4fe-b08d-ac33-ac113fb981ff" [ 991.273331] env[61906]: _type = "HttpNfcLease" [ 991.273331] env[61906]: } obtained for vApp import into resource pool (val){ [ 991.273331] env[61906]: value = "resgroup-9" [ 991.273331] env[61906]: _type = "ResourcePool" [ 991.273331] env[61906]: }. {{(pid=61906) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 991.273670] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the lease: (returnval){ [ 991.273670] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52783e34-a4fe-b08d-ac33-ac113fb981ff" [ 991.273670] env[61906]: _type = "HttpNfcLease" [ 991.273670] env[61906]: } to be ready. {{(pid=61906) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 991.281889] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 991.281889] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52783e34-a4fe-b08d-ac33-ac113fb981ff" [ 991.281889] env[61906]: _type = "HttpNfcLease" [ 991.281889] env[61906]: } is initializing. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 991.288725] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b7bdebfd-3a2f-48b0-b83d-0df6cffe7fd1 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "7264af54-cd10-4507-8ae3-5bb8a5cd127f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.740s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.289984] env[61906]: DEBUG nova.compute.utils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 991.295290] env[61906]: DEBUG nova.compute.manager [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 991.299810] env[61906]: DEBUG nova.network.neutron [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 991.346015] env[61906]: DEBUG nova.policy [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '23ff59c262484422a8badcbca5a103f9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '76ca0b5a4fd24ede8eac4851fc365848', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 991.365624] env[61906]: DEBUG nova.network.neutron [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Updating instance_info_cache with network_info: [{"id": "5cee4cef-6d3a-4191-836c-a40217b08b3c", "address": "fa:16:3e:e2:35:08", "network": {"id": "0a3274f7-785f-45e2-9582-0b611955b024", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-593827400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76ca0b5a4fd24ede8eac4851fc365848", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5cee4cef-6d", "ovs_interfaceid": "5cee4cef-6d3a-4191-836c-a40217b08b3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.508239] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-021fe7bd-3077-42c8-a37e-bbfd03675db6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.517339] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfbef27f-7f80-4870-8ce4-12fa11b8b957 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.551980] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28bc104b-dcee-47d4-a14d-e882a268e94c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.560601] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ca90630-b879-4885-9d8f-310d540441b0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.576714] env[61906]: DEBUG nova.compute.provider_tree [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 991.623039] env[61906]: DEBUG nova.network.neutron [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Successfully created port: d21f5215-2510-4ea0-922d-654e42e6f1d1 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 991.782630] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 991.782630] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52783e34-a4fe-b08d-ac33-ac113fb981ff" [ 991.782630] env[61906]: _type = "HttpNfcLease" [ 991.782630] env[61906]: } is initializing. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 991.803428] env[61906]: DEBUG nova.compute.manager [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 991.826952] env[61906]: DEBUG nova.compute.manager [req-71f5fd9c-3191-40fa-97d2-82dd190489c5 req-b704fb4b-68fb-4069-b1ed-31bd07aeb951 service nova] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Received event network-vif-plugged-5cee4cef-6d3a-4191-836c-a40217b08b3c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.826952] env[61906]: DEBUG oslo_concurrency.lockutils [req-71f5fd9c-3191-40fa-97d2-82dd190489c5 req-b704fb4b-68fb-4069-b1ed-31bd07aeb951 service nova] Acquiring lock "34d22e43-882f-45c2-8259-e77424ce0fb6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.827159] env[61906]: DEBUG oslo_concurrency.lockutils [req-71f5fd9c-3191-40fa-97d2-82dd190489c5 req-b704fb4b-68fb-4069-b1ed-31bd07aeb951 service nova] Lock "34d22e43-882f-45c2-8259-e77424ce0fb6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.830987] env[61906]: DEBUG oslo_concurrency.lockutils [req-71f5fd9c-3191-40fa-97d2-82dd190489c5 req-b704fb4b-68fb-4069-b1ed-31bd07aeb951 service nova] Lock "34d22e43-882f-45c2-8259-e77424ce0fb6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.830987] env[61906]: DEBUG nova.compute.manager [req-71f5fd9c-3191-40fa-97d2-82dd190489c5 req-b704fb4b-68fb-4069-b1ed-31bd07aeb951 service nova] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] No waiting events found dispatching network-vif-plugged-5cee4cef-6d3a-4191-836c-a40217b08b3c {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 991.830987] env[61906]: WARNING nova.compute.manager [req-71f5fd9c-3191-40fa-97d2-82dd190489c5 req-b704fb4b-68fb-4069-b1ed-31bd07aeb951 service nova] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Received unexpected event network-vif-plugged-5cee4cef-6d3a-4191-836c-a40217b08b3c for instance with vm_state building and task_state spawning. [ 991.830987] env[61906]: DEBUG nova.compute.manager [req-71f5fd9c-3191-40fa-97d2-82dd190489c5 req-b704fb4b-68fb-4069-b1ed-31bd07aeb951 service nova] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Received event network-changed-5cee4cef-6d3a-4191-836c-a40217b08b3c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.830987] env[61906]: DEBUG nova.compute.manager [req-71f5fd9c-3191-40fa-97d2-82dd190489c5 req-b704fb4b-68fb-4069-b1ed-31bd07aeb951 service nova] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Refreshing instance network info cache due to event network-changed-5cee4cef-6d3a-4191-836c-a40217b08b3c. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 991.830987] env[61906]: DEBUG oslo_concurrency.lockutils [req-71f5fd9c-3191-40fa-97d2-82dd190489c5 req-b704fb4b-68fb-4069-b1ed-31bd07aeb951 service nova] Acquiring lock "refresh_cache-34d22e43-882f-45c2-8259-e77424ce0fb6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.871017] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Releasing lock "refresh_cache-34d22e43-882f-45c2-8259-e77424ce0fb6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.871017] env[61906]: DEBUG nova.compute.manager [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Instance network_info: |[{"id": "5cee4cef-6d3a-4191-836c-a40217b08b3c", "address": "fa:16:3e:e2:35:08", "network": {"id": "0a3274f7-785f-45e2-9582-0b611955b024", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-593827400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76ca0b5a4fd24ede8eac4851fc365848", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5cee4cef-6d", "ovs_interfaceid": "5cee4cef-6d3a-4191-836c-a40217b08b3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 991.871017] env[61906]: DEBUG oslo_concurrency.lockutils [req-71f5fd9c-3191-40fa-97d2-82dd190489c5 req-b704fb4b-68fb-4069-b1ed-31bd07aeb951 service nova] Acquired lock "refresh_cache-34d22e43-882f-45c2-8259-e77424ce0fb6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.871017] env[61906]: DEBUG nova.network.neutron [req-71f5fd9c-3191-40fa-97d2-82dd190489c5 req-b704fb4b-68fb-4069-b1ed-31bd07aeb951 service nova] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Refreshing network info cache for port 5cee4cef-6d3a-4191-836c-a40217b08b3c {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 991.871017] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e2:35:08', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5cee4cef-6d3a-4191-836c-a40217b08b3c', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 991.878771] env[61906]: DEBUG oslo.service.loopingcall [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 991.879767] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 991.880148] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-837d06a0-b956-40a4-abd0-25606d5cee1e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.903635] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 991.903635] env[61906]: value = "task-1333523" [ 991.903635] env[61906]: _type = "Task" [ 991.903635] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.911927] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333523, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.080100] env[61906]: DEBUG nova.scheduler.client.report [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 992.283397] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 992.283397] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52783e34-a4fe-b08d-ac33-ac113fb981ff" [ 992.283397] env[61906]: _type = "HttpNfcLease" [ 992.283397] env[61906]: } is ready. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 992.283828] env[61906]: DEBUG oslo_vmware.rw_handles [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 992.283828] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52783e34-a4fe-b08d-ac33-ac113fb981ff" [ 992.283828] env[61906]: _type = "HttpNfcLease" [ 992.283828] env[61906]: }. {{(pid=61906) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 992.284640] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ff869e2-d8c1-4219-8e71-6e10c10a177e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.294789] env[61906]: DEBUG oslo_vmware.rw_handles [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b1d20d-abc7-7a0c-ea92-6fcbf6c56184/disk-0.vmdk from lease info. {{(pid=61906) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 992.294789] env[61906]: DEBUG oslo_vmware.rw_handles [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b1d20d-abc7-7a0c-ea92-6fcbf6c56184/disk-0.vmdk. {{(pid=61906) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 992.363804] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-48923617-61a6-4530-bd52-f71ad83614b2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.415182] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333523, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.465445] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "b3e9ab00-8627-4a57-839e-68be5f794b28" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.465702] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "b3e9ab00-8627-4a57-839e-68be5f794b28" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.586991] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.800s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.591114] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 4.551s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.611684] env[61906]: INFO nova.scheduler.client.report [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Deleted allocations for instance f49fb403-fbb4-4e26-8e75-c160d11dea05 [ 992.634664] env[61906]: DEBUG nova.network.neutron [req-71f5fd9c-3191-40fa-97d2-82dd190489c5 req-b704fb4b-68fb-4069-b1ed-31bd07aeb951 service nova] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Updated VIF entry in instance network info cache for port 5cee4cef-6d3a-4191-836c-a40217b08b3c. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 992.635104] env[61906]: DEBUG nova.network.neutron [req-71f5fd9c-3191-40fa-97d2-82dd190489c5 req-b704fb4b-68fb-4069-b1ed-31bd07aeb951 service nova] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Updating instance_info_cache with network_info: [{"id": "5cee4cef-6d3a-4191-836c-a40217b08b3c", "address": "fa:16:3e:e2:35:08", "network": {"id": "0a3274f7-785f-45e2-9582-0b611955b024", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-593827400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76ca0b5a4fd24ede8eac4851fc365848", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5cee4cef-6d", "ovs_interfaceid": "5cee4cef-6d3a-4191-836c-a40217b08b3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.862150] env[61906]: DEBUG nova.compute.manager [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 992.891913] env[61906]: DEBUG nova.virt.hardware [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 992.892210] env[61906]: DEBUG nova.virt.hardware [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 992.892388] env[61906]: DEBUG nova.virt.hardware [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 992.892591] env[61906]: DEBUG nova.virt.hardware [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 992.892783] env[61906]: DEBUG nova.virt.hardware [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 992.892963] env[61906]: DEBUG nova.virt.hardware [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 992.893202] env[61906]: DEBUG nova.virt.hardware [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 992.893372] env[61906]: DEBUG nova.virt.hardware [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 992.893547] env[61906]: DEBUG nova.virt.hardware [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 992.893720] env[61906]: DEBUG nova.virt.hardware [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 992.893934] env[61906]: DEBUG nova.virt.hardware [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 992.894840] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22fb769a-7f09-495f-b9db-413a8d2b428c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.910835] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b31cabd-a4de-4320-87e0-ed5bb2793959 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.923622] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333523, 'name': CreateVM_Task, 'duration_secs': 0.618089} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.936101] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 992.938875] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.939089] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.939424] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 992.939720] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f35da4db-0876-476a-a47b-2696260ab978 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.945959] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 992.945959] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52d9e07b-bb27-7590-d197-5f3c7e4e256c" [ 992.945959] env[61906]: _type = "Task" [ 992.945959] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.958206] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52d9e07b-bb27-7590-d197-5f3c7e4e256c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.968141] env[61906]: DEBUG nova.compute.manager [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 993.102607] env[61906]: INFO nova.compute.claims [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 993.119475] env[61906]: DEBUG oslo_vmware.rw_handles [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Completed reading data from the image iterator. {{(pid=61906) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 993.119858] env[61906]: DEBUG oslo_vmware.rw_handles [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b1d20d-abc7-7a0c-ea92-6fcbf6c56184/disk-0.vmdk. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 993.120947] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d64233b-55fd-4c74-a286-beb28ebf7e6c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.130888] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ac6d023-4c66-443c-9c80-a795cebabf66 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "f49fb403-fbb4-4e26-8e75-c160d11dea05" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.807s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.136096] env[61906]: DEBUG oslo_vmware.rw_handles [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b1d20d-abc7-7a0c-ea92-6fcbf6c56184/disk-0.vmdk is in state: ready. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 993.136096] env[61906]: DEBUG oslo_vmware.rw_handles [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b1d20d-abc7-7a0c-ea92-6fcbf6c56184/disk-0.vmdk. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 993.139614] env[61906]: DEBUG oslo_concurrency.lockutils [req-71f5fd9c-3191-40fa-97d2-82dd190489c5 req-b704fb4b-68fb-4069-b1ed-31bd07aeb951 service nova] Releasing lock "refresh_cache-34d22e43-882f-45c2-8259-e77424ce0fb6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.140074] env[61906]: DEBUG nova.compute.manager [req-71f5fd9c-3191-40fa-97d2-82dd190489c5 req-b704fb4b-68fb-4069-b1ed-31bd07aeb951 service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Received event network-changed-73d0e617-fd5e-4fa2-81c5-710a3beac1a4 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.140179] env[61906]: DEBUG nova.compute.manager [req-71f5fd9c-3191-40fa-97d2-82dd190489c5 req-b704fb4b-68fb-4069-b1ed-31bd07aeb951 service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Refreshing instance network info cache due to event network-changed-73d0e617-fd5e-4fa2-81c5-710a3beac1a4. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 993.140742] env[61906]: DEBUG oslo_concurrency.lockutils [req-71f5fd9c-3191-40fa-97d2-82dd190489c5 req-b704fb4b-68fb-4069-b1ed-31bd07aeb951 service nova] Acquiring lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.140742] env[61906]: DEBUG oslo_concurrency.lockutils [req-71f5fd9c-3191-40fa-97d2-82dd190489c5 req-b704fb4b-68fb-4069-b1ed-31bd07aeb951 service nova] Acquired lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.140742] env[61906]: DEBUG nova.network.neutron [req-71f5fd9c-3191-40fa-97d2-82dd190489c5 req-b704fb4b-68fb-4069-b1ed-31bd07aeb951 service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Refreshing network info cache for port 73d0e617-fd5e-4fa2-81c5-710a3beac1a4 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 993.142133] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-ca1767f7-557d-45fa-8d97-de01a5937f1e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.146404] env[61906]: DEBUG nova.compute.manager [req-2335e3f1-8312-4c52-8d0f-f8a9cf63e5c5 req-1aa89c24-298f-43f1-a335-4b19dd3b6abd service nova] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Received event network-vif-plugged-d21f5215-2510-4ea0-922d-654e42e6f1d1 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.146718] env[61906]: DEBUG oslo_concurrency.lockutils [req-2335e3f1-8312-4c52-8d0f-f8a9cf63e5c5 req-1aa89c24-298f-43f1-a335-4b19dd3b6abd service nova] Acquiring lock "cec87a1e-4176-45bb-a0c7-a594cb399170-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.146983] env[61906]: DEBUG oslo_concurrency.lockutils [req-2335e3f1-8312-4c52-8d0f-f8a9cf63e5c5 req-1aa89c24-298f-43f1-a335-4b19dd3b6abd service nova] Lock "cec87a1e-4176-45bb-a0c7-a594cb399170-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.147234] env[61906]: DEBUG oslo_concurrency.lockutils [req-2335e3f1-8312-4c52-8d0f-f8a9cf63e5c5 req-1aa89c24-298f-43f1-a335-4b19dd3b6abd service nova] Lock "cec87a1e-4176-45bb-a0c7-a594cb399170-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.147531] env[61906]: DEBUG nova.compute.manager [req-2335e3f1-8312-4c52-8d0f-f8a9cf63e5c5 req-1aa89c24-298f-43f1-a335-4b19dd3b6abd service nova] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] No waiting events found dispatching network-vif-plugged-d21f5215-2510-4ea0-922d-654e42e6f1d1 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 993.147766] env[61906]: WARNING nova.compute.manager [req-2335e3f1-8312-4c52-8d0f-f8a9cf63e5c5 req-1aa89c24-298f-43f1-a335-4b19dd3b6abd service nova] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Received unexpected event network-vif-plugged-d21f5215-2510-4ea0-922d-654e42e6f1d1 for instance with vm_state building and task_state spawning. [ 993.237614] env[61906]: DEBUG nova.network.neutron [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Successfully updated port: d21f5215-2510-4ea0-922d-654e42e6f1d1 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 993.291400] env[61906]: DEBUG oslo_concurrency.lockutils [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "0f846f95-2573-4450-9918-a34467d73363" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.291717] env[61906]: DEBUG oslo_concurrency.lockutils [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "0f846f95-2573-4450-9918-a34467d73363" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.293719] env[61906]: DEBUG oslo_concurrency.lockutils [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "0f846f95-2573-4450-9918-a34467d73363-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.293719] env[61906]: DEBUG oslo_concurrency.lockutils [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "0f846f95-2573-4450-9918-a34467d73363-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.293719] env[61906]: DEBUG oslo_concurrency.lockutils [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "0f846f95-2573-4450-9918-a34467d73363-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.296725] env[61906]: INFO nova.compute.manager [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Terminating instance [ 993.299228] env[61906]: DEBUG nova.compute.manager [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 993.299228] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 993.299228] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9da20040-b106-48cf-8f97-2687a896ae13 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.309407] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 993.312320] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f201214a-be77-4b91-b5b5-8dc1ea0447ca {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.316011] env[61906]: DEBUG oslo_vmware.rw_handles [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b1d20d-abc7-7a0c-ea92-6fcbf6c56184/disk-0.vmdk. {{(pid=61906) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 993.316430] env[61906]: INFO nova.virt.vmwareapi.images [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Downloaded image file data 40b9307b-1417-46aa-b262-7664cfccd96f [ 993.316999] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19e001a1-71c6-4f0d-906b-9b0003235187 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.320635] env[61906]: DEBUG oslo_vmware.api [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 993.320635] env[61906]: value = "task-1333524" [ 993.320635] env[61906]: _type = "Task" [ 993.320635] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.335082] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-74ddfeb1-a368-4d89-aad3-727c3085e66c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.339762] env[61906]: DEBUG oslo_vmware.api [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333524, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.387094] env[61906]: INFO nova.virt.vmwareapi.images [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] The imported VM was unregistered [ 993.390357] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Caching image {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 993.390948] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Creating directory with path [datastore2] devstack-image-cache_base/40b9307b-1417-46aa-b262-7664cfccd96f {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 993.391661] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-91401735-6c08-4be0-83f3-64dbcc8dcd76 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.417292] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Created directory with path [datastore2] devstack-image-cache_base/40b9307b-1417-46aa-b262-7664cfccd96f {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 993.417292] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_4f594878-e59c-4af4-a30e-045edbff7e1f/OSTACK_IMG_4f594878-e59c-4af4-a30e-045edbff7e1f.vmdk to [datastore2] devstack-image-cache_base/40b9307b-1417-46aa-b262-7664cfccd96f/40b9307b-1417-46aa-b262-7664cfccd96f.vmdk. {{(pid=61906) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 993.417292] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-81de4800-726f-4c0a-959f-0c8dd998046f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.427526] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 993.427526] env[61906]: value = "task-1333526" [ 993.427526] env[61906]: _type = "Task" [ 993.427526] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.436649] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333526, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.437584] env[61906]: DEBUG nova.network.neutron [req-71f5fd9c-3191-40fa-97d2-82dd190489c5 req-b704fb4b-68fb-4069-b1ed-31bd07aeb951 service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Updated VIF entry in instance network info cache for port 73d0e617-fd5e-4fa2-81c5-710a3beac1a4. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 993.437943] env[61906]: DEBUG nova.network.neutron [req-71f5fd9c-3191-40fa-97d2-82dd190489c5 req-b704fb4b-68fb-4069-b1ed-31bd07aeb951 service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Updating instance_info_cache with network_info: [{"id": "73d0e617-fd5e-4fa2-81c5-710a3beac1a4", "address": "fa:16:3e:66:bf:d9", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73d0e617-fd", "ovs_interfaceid": "73d0e617-fd5e-4fa2-81c5-710a3beac1a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.459445] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52d9e07b-bb27-7590-d197-5f3c7e4e256c, 'name': SearchDatastore_Task, 'duration_secs': 0.023045} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.459739] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.459976] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 993.460230] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.460382] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.460562] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 993.460833] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-57b9f8e2-5d1b-4925-a4d9-2970c96fc420 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.471444] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 993.471579] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 993.472455] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67d47aee-1c13-4452-8963-634bc9b7b2b8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.480297] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 993.480297] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52ebd552-ec5d-a4b2-ae9f-1d23e3ddd9ad" [ 993.480297] env[61906]: _type = "Task" [ 993.480297] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.492825] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52ebd552-ec5d-a4b2-ae9f-1d23e3ddd9ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.496737] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.609296] env[61906]: INFO nova.compute.resource_tracker [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Updating resource usage from migration f4990fe8-e805-4280-8df9-576e6ed39d3b [ 993.740765] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "refresh_cache-cec87a1e-4176-45bb-a0c7-a594cb399170" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.741020] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquired lock "refresh_cache-cec87a1e-4176-45bb-a0c7-a594cb399170" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.741165] env[61906]: DEBUG nova.network.neutron [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 993.780412] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10f65e41-d151-4525-892a-03464413fef2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.789575] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44c187f4-2d4c-4bb2-b2d1-213db3a51a06 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.828127] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6632df40-d01e-47ac-ba8a-639d2a4e3d26 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.839753] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf689b2-1392-4c9d-b2f2-ddb5e2c04660 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.843987] env[61906]: DEBUG oslo_vmware.api [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333524, 'name': PowerOffVM_Task, 'duration_secs': 0.266699} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.844337] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 993.844515] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 993.845245] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e0a90212-9222-4c82-b9b8-4326c3cfa337 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.855656] env[61906]: DEBUG nova.compute.provider_tree [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 993.928662] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 993.928913] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 993.929096] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Deleting the datastore file [datastore2] 0f846f95-2573-4450-9918-a34467d73363 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 993.933237] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-41741b7d-2333-4305-9c9e-d2c691f43f8b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.942410] env[61906]: DEBUG oslo_concurrency.lockutils [req-71f5fd9c-3191-40fa-97d2-82dd190489c5 req-b704fb4b-68fb-4069-b1ed-31bd07aeb951 service nova] Releasing lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.943058] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333526, 'name': MoveVirtualDisk_Task} progress is 12%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.944851] env[61906]: DEBUG oslo_vmware.api [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 993.944851] env[61906]: value = "task-1333528" [ 993.944851] env[61906]: _type = "Task" [ 993.944851] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.955756] env[61906]: DEBUG oslo_vmware.api [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333528, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.994627] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52ebd552-ec5d-a4b2-ae9f-1d23e3ddd9ad, 'name': SearchDatastore_Task, 'duration_secs': 0.015581} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.995669] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f508ba4c-b87f-4f6b-ba15-849e4dfde52e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.002512] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 994.002512] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52ea9c8b-bf46-a150-a52b-4c4bc2ecb2e4" [ 994.002512] env[61906]: _type = "Task" [ 994.002512] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.011587] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52ea9c8b-bf46-a150-a52b-4c4bc2ecb2e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.293397] env[61906]: DEBUG nova.network.neutron [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 994.359558] env[61906]: DEBUG nova.scheduler.client.report [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 994.443484] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333526, 'name': MoveVirtualDisk_Task} progress is 32%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.459117] env[61906]: DEBUG oslo_vmware.api [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333528, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.513684] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52ea9c8b-bf46-a150-a52b-4c4bc2ecb2e4, 'name': SearchDatastore_Task, 'duration_secs': 0.095003} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.514630] env[61906]: DEBUG nova.network.neutron [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Updating instance_info_cache with network_info: [{"id": "d21f5215-2510-4ea0-922d-654e42e6f1d1", "address": "fa:16:3e:7e:9d:92", "network": {"id": "0a3274f7-785f-45e2-9582-0b611955b024", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-593827400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76ca0b5a4fd24ede8eac4851fc365848", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd21f5215-25", "ovs_interfaceid": "d21f5215-2510-4ea0-922d-654e42e6f1d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.519000] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.519000] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 34d22e43-882f-45c2-8259-e77424ce0fb6/34d22e43-882f-45c2-8259-e77424ce0fb6.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 994.519000] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-99dcf07d-6828-4bae-aaa0-850b51c76777 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.526918] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 994.526918] env[61906]: value = "task-1333529" [ 994.526918] env[61906]: _type = "Task" [ 994.526918] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.537987] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333529, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.865063] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.274s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.865347] env[61906]: INFO nova.compute.manager [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Migrating [ 994.872043] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 4.132s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.940775] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333526, 'name': MoveVirtualDisk_Task} progress is 54%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.960836] env[61906]: DEBUG oslo_vmware.api [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333528, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.017687] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Releasing lock "refresh_cache-cec87a1e-4176-45bb-a0c7-a594cb399170" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.018082] env[61906]: DEBUG nova.compute.manager [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Instance network_info: |[{"id": "d21f5215-2510-4ea0-922d-654e42e6f1d1", "address": "fa:16:3e:7e:9d:92", "network": {"id": "0a3274f7-785f-45e2-9582-0b611955b024", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-593827400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76ca0b5a4fd24ede8eac4851fc365848", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd21f5215-25", "ovs_interfaceid": "d21f5215-2510-4ea0-922d-654e42e6f1d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 995.018776] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7e:9d:92', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd21f5215-2510-4ea0-922d-654e42e6f1d1', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 995.027116] env[61906]: DEBUG oslo.service.loopingcall [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 995.027404] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 995.027733] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-677754cc-fdd6-4a7f-849c-8478cc1751dc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.053551] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333529, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.055914] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 995.055914] env[61906]: value = "task-1333530" [ 995.055914] env[61906]: _type = "Task" [ 995.055914] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.064381] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333530, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.168716] env[61906]: DEBUG nova.compute.manager [req-fca40cdb-8a77-45a8-a699-e38065ae23da req-de34c93a-7e1d-4915-a5a3-096909e0cd52 service nova] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Received event network-changed-d21f5215-2510-4ea0-922d-654e42e6f1d1 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 995.168999] env[61906]: DEBUG nova.compute.manager [req-fca40cdb-8a77-45a8-a699-e38065ae23da req-de34c93a-7e1d-4915-a5a3-096909e0cd52 service nova] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Refreshing instance network info cache due to event network-changed-d21f5215-2510-4ea0-922d-654e42e6f1d1. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 995.169314] env[61906]: DEBUG oslo_concurrency.lockutils [req-fca40cdb-8a77-45a8-a699-e38065ae23da req-de34c93a-7e1d-4915-a5a3-096909e0cd52 service nova] Acquiring lock "refresh_cache-cec87a1e-4176-45bb-a0c7-a594cb399170" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.169517] env[61906]: DEBUG oslo_concurrency.lockutils [req-fca40cdb-8a77-45a8-a699-e38065ae23da req-de34c93a-7e1d-4915-a5a3-096909e0cd52 service nova] Acquired lock "refresh_cache-cec87a1e-4176-45bb-a0c7-a594cb399170" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.169771] env[61906]: DEBUG nova.network.neutron [req-fca40cdb-8a77-45a8-a699-e38065ae23da req-de34c93a-7e1d-4915-a5a3-096909e0cd52 service nova] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Refreshing network info cache for port d21f5215-2510-4ea0-922d-654e42e6f1d1 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 995.381748] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "refresh_cache-458a1669-a62b-4313-874a-e49809d5c034" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.382192] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired lock "refresh_cache-458a1669-a62b-4313-874a-e49809d5c034" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.382192] env[61906]: DEBUG nova.network.neutron [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 995.386130] env[61906]: INFO nova.compute.claims [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 995.441853] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333526, 'name': MoveVirtualDisk_Task} progress is 74%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.458160] env[61906]: DEBUG oslo_vmware.api [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333528, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.554230] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333529, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.565944] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333530, 'name': CreateVM_Task} progress is 25%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.703440] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 995.703440] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 995.897470] env[61906]: INFO nova.compute.resource_tracker [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Updating resource usage from migration ae34f086-21e9-4cfa-ba48-060c083baa74 [ 995.939842] env[61906]: DEBUG nova.network.neutron [req-fca40cdb-8a77-45a8-a699-e38065ae23da req-de34c93a-7e1d-4915-a5a3-096909e0cd52 service nova] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Updated VIF entry in instance network info cache for port d21f5215-2510-4ea0-922d-654e42e6f1d1. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 995.940378] env[61906]: DEBUG nova.network.neutron [req-fca40cdb-8a77-45a8-a699-e38065ae23da req-de34c93a-7e1d-4915-a5a3-096909e0cd52 service nova] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Updating instance_info_cache with network_info: [{"id": "d21f5215-2510-4ea0-922d-654e42e6f1d1", "address": "fa:16:3e:7e:9d:92", "network": {"id": "0a3274f7-785f-45e2-9582-0b611955b024", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-593827400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76ca0b5a4fd24ede8eac4851fc365848", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd21f5215-25", "ovs_interfaceid": "d21f5215-2510-4ea0-922d-654e42e6f1d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.949333] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333526, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.961404] env[61906]: DEBUG oslo_vmware.api [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333528, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.862204} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.961682] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 995.961873] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 995.962091] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 995.962278] env[61906]: INFO nova.compute.manager [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 0f846f95-2573-4450-9918-a34467d73363] Took 2.66 seconds to destroy the instance on the hypervisor. [ 995.962532] env[61906]: DEBUG oslo.service.loopingcall [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 995.962734] env[61906]: DEBUG nova.compute.manager [-] [instance: 0f846f95-2573-4450-9918-a34467d73363] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 995.962850] env[61906]: DEBUG nova.network.neutron [-] [instance: 0f846f95-2573-4450-9918-a34467d73363] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 996.055137] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333529, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.065361] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333530, 'name': CreateVM_Task, 'duration_secs': 0.808613} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.065528] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 996.066215] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.066389] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.066709] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 996.069031] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bff34b77-2c0f-4473-80a0-555a1cb3a258 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.074614] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 996.074614] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52a6e25e-968e-f40d-cd28-abd19ae7a41b" [ 996.074614] env[61906]: _type = "Task" [ 996.074614] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.081795] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52a6e25e-968e-f40d-cd28-abd19ae7a41b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.122479] env[61906]: DEBUG nova.network.neutron [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Updating instance_info_cache with network_info: [{"id": "da64f0e8-39b7-4d88-aa6d-4ac0939a0a20", "address": "fa:16:3e:3b:f8:9b", "network": {"id": "af039bdd-ffd2-483d-9655-e7df48936290", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-111531540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f7fc2042e6549cabe98838990f35842", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda64f0e8-39", "ovs_interfaceid": "da64f0e8-39b7-4d88-aa6d-4ac0939a0a20", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.127019] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83c6680a-b1b5-4379-b1fe-b330a3ad171c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.134562] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f181b4de-d736-4a68-b51e-ecfa053b5b1f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.171150] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5532d47f-a88d-445f-bd42-d9b93cd44b06 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.182494] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da6458f0-ed40-4313-8c49-80f6e347c630 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.197478] env[61906]: DEBUG nova.compute.provider_tree [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 996.201817] env[61906]: DEBUG nova.compute.manager [req-296bceb1-d493-46e4-a9bf-b7adf4dec813 req-40f030ba-5649-46b8-865f-1259f4f04206 service nova] [instance: 0f846f95-2573-4450-9918-a34467d73363] Received event network-vif-deleted-12569035-8b00-43d3-bc6d-5a22464fb53d {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 996.201817] env[61906]: INFO nova.compute.manager [req-296bceb1-d493-46e4-a9bf-b7adf4dec813 req-40f030ba-5649-46b8-865f-1259f4f04206 service nova] [instance: 0f846f95-2573-4450-9918-a34467d73363] Neutron deleted interface 12569035-8b00-43d3-bc6d-5a22464fb53d; detaching it from the instance and deleting it from the info cache [ 996.201910] env[61906]: DEBUG nova.network.neutron [req-296bceb1-d493-46e4-a9bf-b7adf4dec813 req-40f030ba-5649-46b8-865f-1259f4f04206 service nova] [instance: 0f846f95-2573-4450-9918-a34467d73363] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.214218] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 996.214449] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Starting heal instance info cache {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 996.441657] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333526, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.444340] env[61906]: DEBUG oslo_concurrency.lockutils [req-fca40cdb-8a77-45a8-a699-e38065ae23da req-de34c93a-7e1d-4915-a5a3-096909e0cd52 service nova] Releasing lock "refresh_cache-cec87a1e-4176-45bb-a0c7-a594cb399170" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.555394] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333529, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.585910] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52a6e25e-968e-f40d-cd28-abd19ae7a41b, 'name': SearchDatastore_Task, 'duration_secs': 0.05494} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.586334] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.586627] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 996.586929] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.587115] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.587309] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 996.587642] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6189e163-5b6d-4d04-bb3e-d6a17a8c6313 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.602327] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 996.602544] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 996.603508] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b07374c3-3bf2-4efe-99a0-244b30e15505 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.610709] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 996.610709] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52935a73-2796-4071-0d3b-fb3fd76d5018" [ 996.610709] env[61906]: _type = "Task" [ 996.610709] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.620187] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52935a73-2796-4071-0d3b-fb3fd76d5018, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.626063] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Releasing lock "refresh_cache-458a1669-a62b-4313-874a-e49809d5c034" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.683348] env[61906]: DEBUG nova.network.neutron [-] [instance: 0f846f95-2573-4450-9918-a34467d73363] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.700463] env[61906]: DEBUG nova.scheduler.client.report [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 996.705521] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cf3e63fc-4c0a-4777-add1-7061b7140f6a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.717680] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8931411e-e650-44ce-88d4-4361983bd349 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.754402] env[61906]: DEBUG nova.compute.manager [req-296bceb1-d493-46e4-a9bf-b7adf4dec813 req-40f030ba-5649-46b8-865f-1259f4f04206 service nova] [instance: 0f846f95-2573-4450-9918-a34467d73363] Detach interface failed, port_id=12569035-8b00-43d3-bc6d-5a22464fb53d, reason: Instance 0f846f95-2573-4450-9918-a34467d73363 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 996.941235] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333526, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.116387} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.941508] env[61906]: INFO nova.virt.vmwareapi.ds_util [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_4f594878-e59c-4af4-a30e-045edbff7e1f/OSTACK_IMG_4f594878-e59c-4af4-a30e-045edbff7e1f.vmdk to [datastore2] devstack-image-cache_base/40b9307b-1417-46aa-b262-7664cfccd96f/40b9307b-1417-46aa-b262-7664cfccd96f.vmdk. [ 996.941701] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Cleaning up location [datastore2] OSTACK_IMG_4f594878-e59c-4af4-a30e-045edbff7e1f {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 996.941875] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_4f594878-e59c-4af4-a30e-045edbff7e1f {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 996.942153] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f6c1c0a0-90d4-4dbd-8f4e-5f95bddd7798 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.949390] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 996.949390] env[61906]: value = "task-1333531" [ 996.949390] env[61906]: _type = "Task" [ 996.949390] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.957953] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333531, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.054641] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333529, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.329531} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.054913] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 34d22e43-882f-45c2-8259-e77424ce0fb6/34d22e43-882f-45c2-8259-e77424ce0fb6.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 997.055148] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 997.055413] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-38b00bc2-c4ec-4c12-993c-889496bedf78 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.064878] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 997.064878] env[61906]: value = "task-1333532" [ 997.064878] env[61906]: _type = "Task" [ 997.064878] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.073675] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333532, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.123345] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52935a73-2796-4071-0d3b-fb3fd76d5018, 'name': SearchDatastore_Task, 'duration_secs': 0.062856} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.124282] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17ab4978-b3c8-4064-b5de-cf2d9a68f055 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.133212] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 997.133212] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]520ef840-55c4-70ee-8091-fac66c6213cc" [ 997.133212] env[61906]: _type = "Task" [ 997.133212] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.142112] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]520ef840-55c4-70ee-8091-fac66c6213cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.184890] env[61906]: INFO nova.compute.manager [-] [instance: 0f846f95-2573-4450-9918-a34467d73363] Took 1.22 seconds to deallocate network for instance. [ 997.206634] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.334s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.206634] env[61906]: INFO nova.compute.manager [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Migrating [ 997.213029] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.716s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.214543] env[61906]: INFO nova.compute.claims [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 997.459721] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333531, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162627} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.460114] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 997.460114] env[61906]: DEBUG oslo_concurrency.lockutils [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Releasing lock "[datastore2] devstack-image-cache_base/40b9307b-1417-46aa-b262-7664cfccd96f/40b9307b-1417-46aa-b262-7664cfccd96f.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.460372] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/40b9307b-1417-46aa-b262-7664cfccd96f/40b9307b-1417-46aa-b262-7664cfccd96f.vmdk to [datastore2] 37124c5c-2021-415b-925b-52e23a3d2973/37124c5c-2021-415b-925b-52e23a3d2973.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 997.460643] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0e11c824-6d11-4d7b-8798-80cdad7405c7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.467979] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 997.467979] env[61906]: value = "task-1333533" [ 997.467979] env[61906]: _type = "Task" [ 997.467979] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.475766] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333533, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.575371] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333532, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070368} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.575654] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 997.576440] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78023e48-40bc-4a69-8cfa-d3cbecd606c8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.598512] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 34d22e43-882f-45c2-8259-e77424ce0fb6/34d22e43-882f-45c2-8259-e77424ce0fb6.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 997.598818] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e39ce773-7e1f-4511-ae1d-d7dc4b450ec6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.621977] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 997.621977] env[61906]: value = "task-1333534" [ 997.621977] env[61906]: _type = "Task" [ 997.621977] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.630187] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333534, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.649197] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]520ef840-55c4-70ee-8091-fac66c6213cc, 'name': SearchDatastore_Task, 'duration_secs': 0.067207} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.649474] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.649785] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] cec87a1e-4176-45bb-a0c7-a594cb399170/cec87a1e-4176-45bb-a0c7-a594cb399170.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 997.650183] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6029c398-8280-42f7-a627-6e02fa01c55c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.658931] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 997.658931] env[61906]: value = "task-1333535" [ 997.658931] env[61906]: _type = "Task" [ 997.658931] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.668896] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333535, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.691958] env[61906]: DEBUG oslo_concurrency.lockutils [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.700689] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f11ba648-cfb1-486e-9ee4-3615336ba251 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "b1159533-c970-49d6-af42-b954b20d92fb" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.701094] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f11ba648-cfb1-486e-9ee4-3615336ba251 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "b1159533-c970-49d6-af42-b954b20d92fb" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.724861] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "refresh_cache-a3e3ad00-6921-4072-8cb1-d80302883513" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.725104] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquired lock "refresh_cache-a3e3ad00-6921-4072-8cb1-d80302883513" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.725185] env[61906]: DEBUG nova.network.neutron [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 997.977814] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333533, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.132687] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333534, 'name': ReconfigVM_Task, 'duration_secs': 0.34096} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.133016] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 34d22e43-882f-45c2-8259-e77424ce0fb6/34d22e43-882f-45c2-8259-e77424ce0fb6.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 998.133702] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-857e58ad-7197-4711-af3a-b71540fe8acf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.141600] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 998.141600] env[61906]: value = "task-1333536" [ 998.141600] env[61906]: _type = "Task" [ 998.141600] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.148275] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4099b88-e863-4188-bb75-a36b28e79295 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.154301] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333536, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.174506] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Updating instance '458a1669-a62b-4313-874a-e49809d5c034' progress to 0 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 998.188435] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333535, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.204711] env[61906]: INFO nova.compute.manager [None req-f11ba648-cfb1-486e-9ee4-3615336ba251 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Detaching volume 92bbddc7-aa3f-4194-b70a-b524783386e8 [ 998.238885] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Didn't find any instances for network info cache update. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 998.239044] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 998.239316] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 998.239559] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 998.239745] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 998.240094] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 998.241504] env[61906]: INFO nova.virt.block_device [None req-f11ba648-cfb1-486e-9ee4-3615336ba251 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Attempting to driver detach volume 92bbddc7-aa3f-4194-b70a-b524783386e8 from mountpoint /dev/sdb [ 998.241808] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-f11ba648-cfb1-486e-9ee4-3615336ba251 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Volume detach. Driver type: vmdk {{(pid=61906) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 998.242116] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-f11ba648-cfb1-486e-9ee4-3615336ba251 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284803', 'volume_id': '92bbddc7-aa3f-4194-b70a-b524783386e8', 'name': 'volume-92bbddc7-aa3f-4194-b70a-b524783386e8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b1159533-c970-49d6-af42-b954b20d92fb', 'attached_at': '', 'detached_at': '', 'volume_id': '92bbddc7-aa3f-4194-b70a-b524783386e8', 'serial': '92bbddc7-aa3f-4194-b70a-b524783386e8'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 998.242472] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 998.242657] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61906) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 998.244210] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1959e70-5a31-46ab-99aa-66200d67fca5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.247564] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 998.282337] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ca01a2c-a988-483b-87ac-9e730a2f108b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.293785] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f24d234b-e1db-4783-b5b9-93506c20dd09 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.323404] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abca1fa9-8733-41b8-86f7-2dd8a3a71ad0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.345200] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-f11ba648-cfb1-486e-9ee4-3615336ba251 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] The volume has not been displaced from its original location: [datastore2] volume-92bbddc7-aa3f-4194-b70a-b524783386e8/volume-92bbddc7-aa3f-4194-b70a-b524783386e8.vmdk. No consolidation needed. {{(pid=61906) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 998.351722] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-f11ba648-cfb1-486e-9ee4-3615336ba251 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Reconfiguring VM instance instance-00000048 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 998.357129] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f0c0d2f-d620-4fb0-8dad-65caaacd5c5d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.378773] env[61906]: DEBUG oslo_vmware.api [None req-f11ba648-cfb1-486e-9ee4-3615336ba251 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 998.378773] env[61906]: value = "task-1333537" [ 998.378773] env[61906]: _type = "Task" [ 998.378773] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.393205] env[61906]: DEBUG oslo_vmware.api [None req-f11ba648-cfb1-486e-9ee4-3615336ba251 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333537, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.483203] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333533, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.504733] env[61906]: DEBUG nova.network.neutron [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Updating instance_info_cache with network_info: [{"id": "e3ce81bf-b27e-47ea-9c48-a17e608a9cd6", "address": "fa:16:3e:03:ef:06", "network": {"id": "ddc08dc9-d08f-4fb9-9bee-1a030ed69f76", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1124067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b33dea0794a48f78b9f519cb269a8c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f1b07b1-e4e5-4842-9090-07fb2c3e124b", "external-id": "nsx-vlan-transportzone-646", "segmentation_id": 646, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3ce81bf-b2", "ovs_interfaceid": "e3ce81bf-b27e-47ea-9c48-a17e608a9cd6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.512569] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a1cee35-2437-4788-9155-106b9724934b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.523090] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a285e8ef-0437-4db8-9bc3-5142c8c8c7ad {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.558474] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3fbf80a-e8a1-4a4d-806d-8634d8cbea49 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.567792] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8679b024-65d2-4ed9-9402-196e35764b0f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.585874] env[61906]: DEBUG nova.compute.provider_tree [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 998.654065] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333536, 'name': Rename_Task, 'duration_secs': 0.380637} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.654065] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 998.654345] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-80de03ba-a4d1-4b42-985d-9d931716bdb6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.664034] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 998.664034] env[61906]: value = "task-1333538" [ 998.664034] env[61906]: _type = "Task" [ 998.664034] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.673371] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333538, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.682836] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333535, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.684643] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 998.684929] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9cb56d25-b5dd-4f6e-bd19-b4a9d06ce37a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.693329] env[61906]: DEBUG oslo_vmware.api [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 998.693329] env[61906]: value = "task-1333539" [ 998.693329] env[61906]: _type = "Task" [ 998.693329] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.702595] env[61906]: DEBUG oslo_vmware.api [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333539, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.756188] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.891659] env[61906]: DEBUG oslo_vmware.api [None req-f11ba648-cfb1-486e-9ee4-3615336ba251 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333537, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.979361] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333533, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.007813] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Releasing lock "refresh_cache-a3e3ad00-6921-4072-8cb1-d80302883513" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.108693] env[61906]: ERROR nova.scheduler.client.report [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [req-5520e9ae-8cf4-4278-a4bc-e05955a54ff4] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID a5b57df7-a16e-44f9-9b2d-23c518860263. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-5520e9ae-8cf4-4278-a4bc-e05955a54ff4"}]} [ 999.127112] env[61906]: DEBUG nova.scheduler.client.report [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Refreshing inventories for resource provider a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 999.144761] env[61906]: DEBUG nova.scheduler.client.report [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Updating ProviderTree inventory for provider a5b57df7-a16e-44f9-9b2d-23c518860263 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 999.145035] env[61906]: DEBUG nova.compute.provider_tree [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 999.158844] env[61906]: DEBUG nova.scheduler.client.report [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Refreshing aggregate associations for resource provider a5b57df7-a16e-44f9-9b2d-23c518860263, aggregates: None {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 999.176029] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333538, 'name': PowerOnVM_Task} progress is 83%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.179266] env[61906]: DEBUG nova.scheduler.client.report [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Refreshing trait associations for resource provider a5b57df7-a16e-44f9-9b2d-23c518860263, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64 {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 999.187821] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333535, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.204946] env[61906]: DEBUG oslo_vmware.api [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333539, 'name': PowerOffVM_Task, 'duration_secs': 0.202943} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.205327] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 999.205525] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Updating instance '458a1669-a62b-4313-874a-e49809d5c034' progress to 17 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 999.393141] env[61906]: DEBUG oslo_vmware.api [None req-f11ba648-cfb1-486e-9ee4-3615336ba251 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333537, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.482588] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333533, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.500314] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-397a5f78-6eea-4ba9-9efc-2cd6ad60d719 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.509457] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32224d3d-6272-42ba-91be-d7d8a5969f56 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.549926] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64128612-557c-47d4-976f-9c73c25870ab {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.559078] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-561f7544-568c-45af-91ad-cb4e2a6fa654 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.575454] env[61906]: DEBUG nova.compute.provider_tree [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 999.676023] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333538, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.688042] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333535, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.950513} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.688386] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] cec87a1e-4176-45bb-a0c7-a594cb399170/cec87a1e-4176-45bb-a0c7-a594cb399170.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 999.688943] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 999.689277] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-be789f49-6b0c-4d5e-bc5a-a624c05d89ac {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.697220] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 999.697220] env[61906]: value = "task-1333540" [ 999.697220] env[61906]: _type = "Task" [ 999.697220] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.707197] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333540, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.712891] env[61906]: DEBUG nova.virt.hardware [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:23Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 999.713206] env[61906]: DEBUG nova.virt.hardware [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 999.713401] env[61906]: DEBUG nova.virt.hardware [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 999.713638] env[61906]: DEBUG nova.virt.hardware [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 999.713892] env[61906]: DEBUG nova.virt.hardware [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 999.714035] env[61906]: DEBUG nova.virt.hardware [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 999.714289] env[61906]: DEBUG nova.virt.hardware [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 999.714447] env[61906]: DEBUG nova.virt.hardware [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 999.714624] env[61906]: DEBUG nova.virt.hardware [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 999.714797] env[61906]: DEBUG nova.virt.hardware [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 999.714981] env[61906]: DEBUG nova.virt.hardware [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 999.720550] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76966daf-574c-4645-97a9-b99bdb6b6100 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.737978] env[61906]: DEBUG oslo_vmware.api [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 999.737978] env[61906]: value = "task-1333541" [ 999.737978] env[61906]: _type = "Task" [ 999.737978] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.752604] env[61906]: DEBUG oslo_vmware.api [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333541, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.892386] env[61906]: DEBUG oslo_vmware.api [None req-f11ba648-cfb1-486e-9ee4-3615336ba251 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333537, 'name': ReconfigVM_Task, 'duration_secs': 1.493664} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.892746] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-f11ba648-cfb1-486e-9ee4-3615336ba251 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Reconfigured VM instance instance-00000048 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 999.897534] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2236988-a4a9-4165-b5ae-17199455db1f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.915105] env[61906]: DEBUG oslo_vmware.api [None req-f11ba648-cfb1-486e-9ee4-3615336ba251 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 999.915105] env[61906]: value = "task-1333542" [ 999.915105] env[61906]: _type = "Task" [ 999.915105] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.925411] env[61906]: DEBUG oslo_vmware.api [None req-f11ba648-cfb1-486e-9ee4-3615336ba251 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333542, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.982668] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333533, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.107914] env[61906]: DEBUG nova.scheduler.client.report [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Updated inventory for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with generation 114 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1000.108321] env[61906]: DEBUG nova.compute.provider_tree [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Updating resource provider a5b57df7-a16e-44f9-9b2d-23c518860263 generation from 114 to 115 during operation: update_inventory {{(pid=61906) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1000.108553] env[61906]: DEBUG nova.compute.provider_tree [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1000.176203] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333538, 'name': PowerOnVM_Task, 'duration_secs': 1.06789} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.176627] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1000.176852] env[61906]: INFO nova.compute.manager [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Took 10.57 seconds to spawn the instance on the hypervisor. [ 1000.177030] env[61906]: DEBUG nova.compute.manager [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1000.177978] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c67069df-e085-4896-becf-d1506b15c308 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.207765] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333540, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088704} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.208250] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1000.208980] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05ab2e77-1447-4002-b655-588c547dfb0e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.232386] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] cec87a1e-4176-45bb-a0c7-a594cb399170/cec87a1e-4176-45bb-a0c7-a594cb399170.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1000.233470] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-392a2aed-a363-4073-aca5-2b71f038d510 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.258320] env[61906]: DEBUG oslo_vmware.api [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333541, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.259615] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 1000.259615] env[61906]: value = "task-1333543" [ 1000.259615] env[61906]: _type = "Task" [ 1000.259615] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.270021] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333543, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.430052] env[61906]: DEBUG oslo_vmware.api [None req-f11ba648-cfb1-486e-9ee4-3615336ba251 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333542, 'name': ReconfigVM_Task, 'duration_secs': 0.458661} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.430052] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-f11ba648-cfb1-486e-9ee4-3615336ba251 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284803', 'volume_id': '92bbddc7-aa3f-4194-b70a-b524783386e8', 'name': 'volume-92bbddc7-aa3f-4194-b70a-b524783386e8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b1159533-c970-49d6-af42-b954b20d92fb', 'attached_at': '', 'detached_at': '', 'volume_id': '92bbddc7-aa3f-4194-b70a-b524783386e8', 'serial': '92bbddc7-aa3f-4194-b70a-b524783386e8'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1000.482263] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333533, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.530921] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b66f600-057a-40e5-a846-7ac9e6967526 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.551719] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Updating instance 'a3e3ad00-6921-4072-8cb1-d80302883513' progress to 0 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1000.613633] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.400s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.614221] env[61906]: DEBUG nova.compute.manager [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1000.617185] env[61906]: DEBUG oslo_concurrency.lockutils [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.925s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.617559] env[61906]: DEBUG nova.objects.instance [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lazy-loading 'resources' on Instance uuid 0f846f95-2573-4450-9918-a34467d73363 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1000.698031] env[61906]: INFO nova.compute.manager [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Took 15.44 seconds to build instance. [ 1000.758297] env[61906]: DEBUG oslo_vmware.api [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333541, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.768482] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333543, 'name': ReconfigVM_Task, 'duration_secs': 0.43935} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.768745] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Reconfigured VM instance instance-0000005d to attach disk [datastore2] cec87a1e-4176-45bb-a0c7-a594cb399170/cec87a1e-4176-45bb-a0c7-a594cb399170.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1000.769369] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dda1fb7a-2bc6-439d-a9c3-4dc242bca96d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.776204] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 1000.776204] env[61906]: value = "task-1333544" [ 1000.776204] env[61906]: _type = "Task" [ 1000.776204] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.784277] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333544, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.869741] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "150bf47d-7da9-4e13-ad00-7a8a25b9504e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.870093] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "150bf47d-7da9-4e13-ad00-7a8a25b9504e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.870375] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "150bf47d-7da9-4e13-ad00-7a8a25b9504e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.870576] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "150bf47d-7da9-4e13-ad00-7a8a25b9504e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.870778] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "150bf47d-7da9-4e13-ad00-7a8a25b9504e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.873107] env[61906]: INFO nova.compute.manager [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Terminating instance [ 1000.875236] env[61906]: DEBUG nova.compute.manager [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1000.875544] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1000.876848] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71085e29-bb2c-41eb-b378-861221ce4ea2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.886907] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1000.887194] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3f5a7261-4110-4d86-a30a-83cc36e64e5f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.894275] env[61906]: DEBUG oslo_vmware.api [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1000.894275] env[61906]: value = "task-1333545" [ 1000.894275] env[61906]: _type = "Task" [ 1000.894275] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.902397] env[61906]: DEBUG oslo_vmware.api [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333545, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.977846] env[61906]: DEBUG nova.objects.instance [None req-f11ba648-cfb1-486e-9ee4-3615336ba251 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lazy-loading 'flavor' on Instance uuid b1159533-c970-49d6-af42-b954b20d92fb {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1000.982518] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333533, 'name': CopyVirtualDisk_Task, 'duration_secs': 3.046072} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.983665] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/40b9307b-1417-46aa-b262-7664cfccd96f/40b9307b-1417-46aa-b262-7664cfccd96f.vmdk to [datastore2] 37124c5c-2021-415b-925b-52e23a3d2973/37124c5c-2021-415b-925b-52e23a3d2973.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1000.983665] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d19e7500-c5c5-438c-b1f9-4d9371cfe04c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.006338] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] 37124c5c-2021-415b-925b-52e23a3d2973/37124c5c-2021-415b-925b-52e23a3d2973.vmdk or device None with type streamOptimized {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1001.007377] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d7706e82-bb15-4f1c-a9b5-b44ff8fe52b5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.045728] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 1001.045728] env[61906]: value = "task-1333546" [ 1001.045728] env[61906]: _type = "Task" [ 1001.045728] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.054652] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333546, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.057910] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1001.058214] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-58ae1c34-611e-4549-844d-46840ed89cce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.065393] env[61906]: DEBUG oslo_vmware.api [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 1001.065393] env[61906]: value = "task-1333547" [ 1001.065393] env[61906]: _type = "Task" [ 1001.065393] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.074651] env[61906]: DEBUG oslo_vmware.api [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333547, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.125545] env[61906]: DEBUG nova.compute.utils [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1001.128174] env[61906]: DEBUG nova.compute.manager [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1001.128275] env[61906]: DEBUG nova.network.neutron [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1001.186519] env[61906]: DEBUG nova.policy [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0de9abfd085f4a43a878da8448615bc4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e36ff7a500ef444cbf5a168c5a48208b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1001.198914] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "34d22e43-882f-45c2-8259-e77424ce0fb6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.956s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.260694] env[61906]: DEBUG oslo_vmware.api [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333541, 'name': ReconfigVM_Task, 'duration_secs': 1.308311} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.263508] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Updating instance '458a1669-a62b-4313-874a-e49809d5c034' progress to 33 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1001.287546] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333544, 'name': Rename_Task, 'duration_secs': 0.146979} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.290509] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1001.291211] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cceffe82-c840-4111-8ff8-7bb2187c2d74 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.299253] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 1001.299253] env[61906]: value = "task-1333548" [ 1001.299253] env[61906]: _type = "Task" [ 1001.299253] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.311467] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333548, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.374217] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4325235-f466-4771-9e4d-e35b38427bce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.389898] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-045093a5-6e52-4a53-80ee-68562c6817d3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.425391] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70b7d2bd-0e6b-47dc-918f-92ef5f6a348e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.431012] env[61906]: DEBUG oslo_vmware.api [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333545, 'name': PowerOffVM_Task, 'duration_secs': 0.189675} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.432637] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1001.432916] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1001.437298] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-314cc6fd-dec4-40ad-9ddf-1e9d70277c2e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.440033] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03810279-1b40-4828-b0ef-08137aaf567c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.456698] env[61906]: DEBUG nova.compute.provider_tree [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1001.512260] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1001.512493] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1001.513176] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Deleting the datastore file [datastore1] 150bf47d-7da9-4e13-ad00-7a8a25b9504e {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1001.513753] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2608847a-9017-415b-a3b7-01094c4adad0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.521718] env[61906]: DEBUG oslo_vmware.api [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1001.521718] env[61906]: value = "task-1333550" [ 1001.521718] env[61906]: _type = "Task" [ 1001.521718] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.536487] env[61906]: DEBUG oslo_vmware.api [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333550, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.561173] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333546, 'name': ReconfigVM_Task, 'duration_secs': 0.386334} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.561492] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Reconfigured VM instance instance-0000005b to attach disk [datastore2] 37124c5c-2021-415b-925b-52e23a3d2973/37124c5c-2021-415b-925b-52e23a3d2973.vmdk or device None with type streamOptimized {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1001.562189] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-043ecf0a-1f21-4434-9514-53132b899ec6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.571721] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 1001.571721] env[61906]: value = "task-1333551" [ 1001.571721] env[61906]: _type = "Task" [ 1001.571721] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.575145] env[61906]: DEBUG oslo_vmware.api [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333547, 'name': PowerOffVM_Task, 'duration_secs': 0.174998} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.578408] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1001.578679] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Updating instance 'a3e3ad00-6921-4072-8cb1-d80302883513' progress to 17 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1001.585020] env[61906]: DEBUG nova.network.neutron [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Successfully created port: e0244018-d931-4a96-a25e-228b02a88f56 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1001.591032] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333551, 'name': Rename_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.631413] env[61906]: DEBUG nova.compute.manager [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1001.769993] env[61906]: DEBUG nova.virt.hardware [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1001.770318] env[61906]: DEBUG nova.virt.hardware [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1001.770484] env[61906]: DEBUG nova.virt.hardware [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1001.770674] env[61906]: DEBUG nova.virt.hardware [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1001.770917] env[61906]: DEBUG nova.virt.hardware [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1001.771169] env[61906]: DEBUG nova.virt.hardware [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1001.771517] env[61906]: DEBUG nova.virt.hardware [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1001.771710] env[61906]: DEBUG nova.virt.hardware [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1001.771928] env[61906]: DEBUG nova.virt.hardware [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1001.772132] env[61906]: DEBUG nova.virt.hardware [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1001.772318] env[61906]: DEBUG nova.virt.hardware [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1001.777653] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Reconfiguring VM instance instance-00000058 to detach disk 2000 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1001.777970] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6810af13-eff0-4acd-b299-813b4b01e4fa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.798078] env[61906]: DEBUG oslo_vmware.api [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 1001.798078] env[61906]: value = "task-1333552" [ 1001.798078] env[61906]: _type = "Task" [ 1001.798078] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.810049] env[61906]: DEBUG oslo_vmware.api [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333552, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.813032] env[61906]: DEBUG oslo_vmware.api [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333548, 'name': PowerOnVM_Task, 'duration_secs': 0.462217} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.813299] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1001.813515] env[61906]: INFO nova.compute.manager [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Took 8.95 seconds to spawn the instance on the hypervisor. [ 1001.813691] env[61906]: DEBUG nova.compute.manager [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1001.814459] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d43055f1-6b6f-4b31-9b39-9fd2bbd92808 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.940553] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c5db4317-d69d-4425-8fe1-14aa93f877ed tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "b1159533-c970-49d6-af42-b954b20d92fb" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.959916] env[61906]: DEBUG nova.scheduler.client.report [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1001.987715] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f11ba648-cfb1-486e-9ee4-3615336ba251 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "b1159533-c970-49d6-af42-b954b20d92fb" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.287s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.988813] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c5db4317-d69d-4425-8fe1-14aa93f877ed tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "b1159533-c970-49d6-af42-b954b20d92fb" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.048s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.989067] env[61906]: DEBUG nova.compute.manager [None req-c5db4317-d69d-4425-8fe1-14aa93f877ed tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1001.990111] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1e2fd25-a737-4818-ba48-ed088e6f07aa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.998477] env[61906]: DEBUG nova.compute.manager [None req-c5db4317-d69d-4425-8fe1-14aa93f877ed tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61906) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1001.999156] env[61906]: DEBUG nova.objects.instance [None req-c5db4317-d69d-4425-8fe1-14aa93f877ed tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lazy-loading 'flavor' on Instance uuid b1159533-c970-49d6-af42-b954b20d92fb {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1002.031743] env[61906]: DEBUG oslo_vmware.api [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333550, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154759} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.032072] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1002.032313] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1002.032537] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1002.032754] env[61906]: INFO nova.compute.manager [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1002.033165] env[61906]: DEBUG oslo.service.loopingcall [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1002.033354] env[61906]: DEBUG nova.compute.manager [-] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1002.033456] env[61906]: DEBUG nova.network.neutron [-] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1002.084592] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333551, 'name': Rename_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.086585] env[61906]: DEBUG nova.virt.hardware [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:23Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1002.086815] env[61906]: DEBUG nova.virt.hardware [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1002.086982] env[61906]: DEBUG nova.virt.hardware [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1002.087197] env[61906]: DEBUG nova.virt.hardware [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1002.087352] env[61906]: DEBUG nova.virt.hardware [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1002.087506] env[61906]: DEBUG nova.virt.hardware [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1002.087716] env[61906]: DEBUG nova.virt.hardware [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1002.087881] env[61906]: DEBUG nova.virt.hardware [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1002.088067] env[61906]: DEBUG nova.virt.hardware [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1002.088239] env[61906]: DEBUG nova.virt.hardware [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1002.088413] env[61906]: DEBUG nova.virt.hardware [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1002.093544] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4d4afd7-eb30-41c6-83cb-02fc1a2c9df4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.109166] env[61906]: DEBUG oslo_vmware.api [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 1002.109166] env[61906]: value = "task-1333553" [ 1002.109166] env[61906]: _type = "Task" [ 1002.109166] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.122860] env[61906]: DEBUG oslo_vmware.api [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333553, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.312855] env[61906]: DEBUG oslo_vmware.api [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333552, 'name': ReconfigVM_Task, 'duration_secs': 0.448664} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.313250] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Reconfigured VM instance instance-00000058 to detach disk 2000 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1002.314141] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70f3d94e-158d-45ed-a1f7-df8c90fee040 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.343656] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 458a1669-a62b-4313-874a-e49809d5c034/458a1669-a62b-4313-874a-e49809d5c034.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1002.345684] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-65e6922b-e10d-4bbe-8124-1d132d20a224 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.358945] env[61906]: INFO nova.compute.manager [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Took 17.08 seconds to build instance. [ 1002.367033] env[61906]: DEBUG oslo_vmware.api [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 1002.367033] env[61906]: value = "task-1333554" [ 1002.367033] env[61906]: _type = "Task" [ 1002.367033] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.382485] env[61906]: DEBUG oslo_vmware.api [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333554, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.467019] env[61906]: DEBUG oslo_concurrency.lockutils [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.848s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.467633] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 3.712s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.467992] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.468311] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61906) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1002.469683] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16e77008-cfb5-4058-9e5c-ce6a7c9cf250 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.481084] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b03860b-8bbf-43b8-a797-6fbf9894a6ef {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.497976] env[61906]: INFO nova.scheduler.client.report [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Deleted allocations for instance 0f846f95-2573-4450-9918-a34467d73363 [ 1002.499900] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2e60ca4-ce89-458f-856f-5ef199165ae9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.512466] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5db4317-d69d-4425-8fe1-14aa93f877ed tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1002.512764] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b0a5db8a-687d-4bfa-b7f3-5f78e82a2b46 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.517598] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c133f9e1-8e53-42b2-a7cd-f2706bf46564 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.524078] env[61906]: DEBUG oslo_vmware.api [None req-c5db4317-d69d-4425-8fe1-14aa93f877ed tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 1002.524078] env[61906]: value = "task-1333555" [ 1002.524078] env[61906]: _type = "Task" [ 1002.524078] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.555529] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180313MB free_disk=130GB free_vcpus=48 pci_devices=None {{(pid=61906) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1002.556015] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.556015] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.563254] env[61906]: DEBUG oslo_vmware.api [None req-c5db4317-d69d-4425-8fe1-14aa93f877ed tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333555, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.584790] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333551, 'name': Rename_Task, 'duration_secs': 1.006951} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.585115] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1002.585448] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9eb7dca7-6d94-4bfa-81f8-50049ae7bae4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.593786] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 1002.593786] env[61906]: value = "task-1333556" [ 1002.593786] env[61906]: _type = "Task" [ 1002.593786] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.603194] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333556, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.620315] env[61906]: DEBUG oslo_vmware.api [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333553, 'name': ReconfigVM_Task, 'duration_secs': 0.481547} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.620664] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Updating instance 'a3e3ad00-6921-4072-8cb1-d80302883513' progress to 33 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1002.643206] env[61906]: DEBUG nova.compute.manager [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1002.673248] env[61906]: DEBUG nova.virt.hardware [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1002.673518] env[61906]: DEBUG nova.virt.hardware [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1002.673682] env[61906]: DEBUG nova.virt.hardware [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1002.673869] env[61906]: DEBUG nova.virt.hardware [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1002.674033] env[61906]: DEBUG nova.virt.hardware [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1002.674195] env[61906]: DEBUG nova.virt.hardware [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1002.674409] env[61906]: DEBUG nova.virt.hardware [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1002.674668] env[61906]: DEBUG nova.virt.hardware [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1002.674970] env[61906]: DEBUG nova.virt.hardware [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1002.675287] env[61906]: DEBUG nova.virt.hardware [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1002.675594] env[61906]: DEBUG nova.virt.hardware [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1002.677267] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9af4d50-e0a4-408f-a37c-807bcf62c32d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.686863] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f42544a3-2ff2-44ca-830d-8ea8e277b062 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.704800] env[61906]: DEBUG nova.compute.manager [req-b6aeaf0a-3d89-4f60-ad37-175a0ece1cf3 req-ac8c29b0-fa0e-4e4e-af40-8c06e7f2ba5d service nova] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Received event network-vif-deleted-5c99ae7c-41d1-48de-a53f-fc8d1a6a4ca2 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1002.705030] env[61906]: INFO nova.compute.manager [req-b6aeaf0a-3d89-4f60-ad37-175a0ece1cf3 req-ac8c29b0-fa0e-4e4e-af40-8c06e7f2ba5d service nova] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Neutron deleted interface 5c99ae7c-41d1-48de-a53f-fc8d1a6a4ca2; detaching it from the instance and deleting it from the info cache [ 1002.705214] env[61906]: DEBUG nova.network.neutron [req-b6aeaf0a-3d89-4f60-ad37-175a0ece1cf3 req-ac8c29b0-fa0e-4e4e-af40-8c06e7f2ba5d service nova] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.861068] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9948d84-5686-4eab-a538-2d0849ebce5a tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "cec87a1e-4176-45bb-a0c7-a594cb399170" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.594s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.878227] env[61906]: DEBUG oslo_vmware.api [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333554, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.012254] env[61906]: DEBUG oslo_concurrency.lockutils [None req-724c1ea8-906e-4fc0-8333-ee5f4d1b22f8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "0f846f95-2573-4450-9918-a34467d73363" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.720s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.037925] env[61906]: DEBUG oslo_vmware.api [None req-c5db4317-d69d-4425-8fe1-14aa93f877ed tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333555, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.092008] env[61906]: DEBUG nova.compute.manager [req-7f2bcb5a-af89-4b8a-84bb-593e53cff070 req-35166ad6-343e-402d-b6eb-fc37e3720ed8 service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Received event network-vif-plugged-e0244018-d931-4a96-a25e-228b02a88f56 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1003.092396] env[61906]: DEBUG oslo_concurrency.lockutils [req-7f2bcb5a-af89-4b8a-84bb-593e53cff070 req-35166ad6-343e-402d-b6eb-fc37e3720ed8 service nova] Acquiring lock "b3e9ab00-8627-4a57-839e-68be5f794b28-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.092745] env[61906]: DEBUG oslo_concurrency.lockutils [req-7f2bcb5a-af89-4b8a-84bb-593e53cff070 req-35166ad6-343e-402d-b6eb-fc37e3720ed8 service nova] Lock "b3e9ab00-8627-4a57-839e-68be5f794b28-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.093148] env[61906]: DEBUG oslo_concurrency.lockutils [req-7f2bcb5a-af89-4b8a-84bb-593e53cff070 req-35166ad6-343e-402d-b6eb-fc37e3720ed8 service nova] Lock "b3e9ab00-8627-4a57-839e-68be5f794b28-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.093370] env[61906]: DEBUG nova.compute.manager [req-7f2bcb5a-af89-4b8a-84bb-593e53cff070 req-35166ad6-343e-402d-b6eb-fc37e3720ed8 service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] No waiting events found dispatching network-vif-plugged-e0244018-d931-4a96-a25e-228b02a88f56 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1003.093666] env[61906]: WARNING nova.compute.manager [req-7f2bcb5a-af89-4b8a-84bb-593e53cff070 req-35166ad6-343e-402d-b6eb-fc37e3720ed8 service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Received unexpected event network-vif-plugged-e0244018-d931-4a96-a25e-228b02a88f56 for instance with vm_state building and task_state spawning. [ 1003.104208] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333556, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.128048] env[61906]: DEBUG nova.virt.hardware [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1003.128329] env[61906]: DEBUG nova.virt.hardware [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1003.128492] env[61906]: DEBUG nova.virt.hardware [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1003.128677] env[61906]: DEBUG nova.virt.hardware [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1003.128828] env[61906]: DEBUG nova.virt.hardware [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1003.128983] env[61906]: DEBUG nova.virt.hardware [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1003.129210] env[61906]: DEBUG nova.virt.hardware [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1003.129373] env[61906]: DEBUG nova.virt.hardware [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1003.129540] env[61906]: DEBUG nova.virt.hardware [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1003.129809] env[61906]: DEBUG nova.virt.hardware [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1003.130022] env[61906]: DEBUG nova.virt.hardware [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1003.135560] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Reconfiguring VM instance instance-00000059 to detach disk 2000 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1003.136220] env[61906]: DEBUG nova.network.neutron [-] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.137417] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46c2a8da-b823-423d-8580-95893d9e211c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.158531] env[61906]: DEBUG oslo_vmware.api [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 1003.158531] env[61906]: value = "task-1333557" [ 1003.158531] env[61906]: _type = "Task" [ 1003.158531] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.167434] env[61906]: DEBUG oslo_vmware.api [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333557, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.192017] env[61906]: DEBUG nova.network.neutron [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Successfully updated port: e0244018-d931-4a96-a25e-228b02a88f56 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1003.208726] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c55755c5-e91a-4ed1-b85f-8c987b90c19b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.223889] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c5bfac5-d381-4fa3-b51c-db59653a7fdd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.263661] env[61906]: DEBUG nova.compute.manager [req-b6aeaf0a-3d89-4f60-ad37-175a0ece1cf3 req-ac8c29b0-fa0e-4e4e-af40-8c06e7f2ba5d service nova] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Detach interface failed, port_id=5c99ae7c-41d1-48de-a53f-fc8d1a6a4ca2, reason: Instance 150bf47d-7da9-4e13-ad00-7a8a25b9504e could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1003.316975] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "34d22e43-882f-45c2-8259-e77424ce0fb6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.317445] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "34d22e43-882f-45c2-8259-e77424ce0fb6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.317669] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "34d22e43-882f-45c2-8259-e77424ce0fb6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.317864] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "34d22e43-882f-45c2-8259-e77424ce0fb6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.318064] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "34d22e43-882f-45c2-8259-e77424ce0fb6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.321388] env[61906]: INFO nova.compute.manager [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Terminating instance [ 1003.323901] env[61906]: DEBUG nova.compute.manager [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1003.324112] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1003.324931] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52b79581-2bd2-40b5-a444-8a2dd5efeaab {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.333896] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1003.334236] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8b80e13d-0a13-478c-b7f5-341515ca5163 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.341908] env[61906]: DEBUG oslo_vmware.api [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 1003.341908] env[61906]: value = "task-1333558" [ 1003.341908] env[61906]: _type = "Task" [ 1003.341908] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.350315] env[61906]: DEBUG oslo_vmware.api [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333558, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.378463] env[61906]: DEBUG oslo_vmware.api [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333554, 'name': ReconfigVM_Task, 'duration_secs': 0.551089} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.378748] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 458a1669-a62b-4313-874a-e49809d5c034/458a1669-a62b-4313-874a-e49809d5c034.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1003.379043] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Updating instance '458a1669-a62b-4313-874a-e49809d5c034' progress to 50 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1003.386078] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "cec87a1e-4176-45bb-a0c7-a594cb399170" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.386264] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "cec87a1e-4176-45bb-a0c7-a594cb399170" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.386518] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "cec87a1e-4176-45bb-a0c7-a594cb399170-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.386981] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "cec87a1e-4176-45bb-a0c7-a594cb399170-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.387208] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "cec87a1e-4176-45bb-a0c7-a594cb399170-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.389869] env[61906]: INFO nova.compute.manager [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Terminating instance [ 1003.391565] env[61906]: DEBUG nova.compute.manager [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1003.391749] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1003.392577] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cc791e2-ee87-43c7-ade4-7aa2a845d20d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.401236] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1003.401509] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-225384ea-ed43-4003-a539-17f6ae7426d3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.409102] env[61906]: DEBUG oslo_vmware.api [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 1003.409102] env[61906]: value = "task-1333559" [ 1003.409102] env[61906]: _type = "Task" [ 1003.409102] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.421938] env[61906]: DEBUG oslo_vmware.api [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333559, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.543438] env[61906]: DEBUG oslo_vmware.api [None req-c5db4317-d69d-4425-8fe1-14aa93f877ed tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333555, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.566550] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Applying migration context for instance 458a1669-a62b-4313-874a-e49809d5c034 as it has an incoming, in-progress migration f4990fe8-e805-4280-8df9-576e6ed39d3b. Migration status is migrating {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1003.566883] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Applying migration context for instance a3e3ad00-6921-4072-8cb1-d80302883513 as it has an incoming, in-progress migration ae34f086-21e9-4cfa-ba48-060c083baa74. Migration status is migrating {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1003.569415] env[61906]: INFO nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Updating resource usage from migration f4990fe8-e805-4280-8df9-576e6ed39d3b [ 1003.569830] env[61906]: INFO nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Updating resource usage from migration ae34f086-21e9-4cfa-ba48-060c083baa74 [ 1003.592914] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 7504dfa6-bedc-4701-b4fc-60e19e742276 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1003.593224] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance b1159533-c970-49d6-af42-b954b20d92fb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1003.593347] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 1a335019-915b-4065-b22f-d2f0af728d6f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1003.593484] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 150bf47d-7da9-4e13-ad00-7a8a25b9504e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1003.593654] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 03d3fb1b-0e46-4544-b01d-498a2baf3b45 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1003.593727] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 7264af54-cd10-4507-8ae3-5bb8a5cd127f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1003.593847] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 37124c5c-2021-415b-925b-52e23a3d2973 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1003.594503] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 34d22e43-882f-45c2-8259-e77424ce0fb6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1003.594503] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance cec87a1e-4176-45bb-a0c7-a594cb399170 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1003.594503] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Migration f4990fe8-e805-4280-8df9-576e6ed39d3b is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1003.594503] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 458a1669-a62b-4313-874a-e49809d5c034 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1003.594503] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Migration ae34f086-21e9-4cfa-ba48-060c083baa74 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1003.594780] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance a3e3ad00-6921-4072-8cb1-d80302883513 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1003.594780] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance b3e9ab00-8627-4a57-839e-68be5f794b28 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1003.595077] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Total usable vcpus: 48, total allocated vcpus: 14 {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1003.595245] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3328MB phys_disk=200GB used_disk=14GB total_vcpus=48 used_vcpus=14 pci_stats=[] {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1003.607624] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333556, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.652360] env[61906]: INFO nova.compute.manager [-] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Took 1.62 seconds to deallocate network for instance. [ 1003.673204] env[61906]: DEBUG oslo_vmware.api [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333557, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.694668] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "refresh_cache-b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.694880] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired lock "refresh_cache-b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.695146] env[61906]: DEBUG nova.network.neutron [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1003.796343] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db323d9-e328-4483-8cde-c7b0c458e4b2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.805621] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-132f63f6-7945-4f0d-89bc-d76bafad9edc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.836929] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef9cb03c-ad35-41e2-8514-fbcb784ced9b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.847343] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27b8f002-8541-4329-b073-4aadf5d8853d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.854111] env[61906]: DEBUG oslo_vmware.api [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333558, 'name': PowerOffVM_Task, 'duration_secs': 0.1766} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.854849] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1003.854955] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1003.855202] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-126829a7-089a-4df0-896a-806915b502a7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.864110] env[61906]: DEBUG nova.compute.provider_tree [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1003.885966] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81e4afdf-b876-4698-a172-a5d74e8d04ad {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.907278] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b29c67d4-f975-4111-a495-eb13a81ead52 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.928868] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Updating instance '458a1669-a62b-4313-874a-e49809d5c034' progress to 67 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1003.937134] env[61906]: DEBUG oslo_vmware.api [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333559, 'name': PowerOffVM_Task, 'duration_secs': 0.187662} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.937400] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1003.937863] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1003.937863] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Deleting the datastore file [datastore2] 34d22e43-882f-45c2-8259-e77424ce0fb6 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1003.938075] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1003.938232] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1003.938455] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c05a1be9-b531-4500-b638-6626d0fe1433 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.940295] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-59e48edb-8e77-4d0a-a603-2b391c78dfca {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.947105] env[61906]: DEBUG oslo_vmware.api [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 1003.947105] env[61906]: value = "task-1333561" [ 1003.947105] env[61906]: _type = "Task" [ 1003.947105] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.955664] env[61906]: DEBUG oslo_vmware.api [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333561, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.009906] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1004.009906] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1004.009906] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Deleting the datastore file [datastore2] cec87a1e-4176-45bb-a0c7-a594cb399170 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1004.010045] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c70ab1a8-89ab-4ffd-94af-a817d9ae3351 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.017727] env[61906]: DEBUG oslo_vmware.api [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for the task: (returnval){ [ 1004.017727] env[61906]: value = "task-1333563" [ 1004.017727] env[61906]: _type = "Task" [ 1004.017727] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.028724] env[61906]: DEBUG oslo_vmware.api [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333563, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.038199] env[61906]: DEBUG oslo_vmware.api [None req-c5db4317-d69d-4425-8fe1-14aa93f877ed tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333555, 'name': PowerOffVM_Task, 'duration_secs': 1.194105} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.038531] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5db4317-d69d-4425-8fe1-14aa93f877ed tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1004.038790] env[61906]: DEBUG nova.compute.manager [None req-c5db4317-d69d-4425-8fe1-14aa93f877ed tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1004.039644] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8042a094-a5e1-4be2-8a8a-4912388fa471 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.091276] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "df605d77-7766-47b3-bd27-97361ff8ae4d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.091471] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "df605d77-7766-47b3-bd27-97361ff8ae4d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.109229] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333556, 'name': PowerOnVM_Task} progress is 90%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.165685] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.171883] env[61906]: DEBUG oslo_vmware.api [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333557, 'name': ReconfigVM_Task, 'duration_secs': 0.52227} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.172165] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Reconfigured VM instance instance-00000059 to detach disk 2000 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1004.173026] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef6c588b-c566-46ce-bfd9-7a84448c375e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.197779] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] a3e3ad00-6921-4072-8cb1-d80302883513/a3e3ad00-6921-4072-8cb1-d80302883513.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1004.199730] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ad7ffa7-77e9-446b-bc9b-512c2c1b2d3b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.219542] env[61906]: DEBUG oslo_vmware.api [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 1004.219542] env[61906]: value = "task-1333564" [ 1004.219542] env[61906]: _type = "Task" [ 1004.219542] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.231123] env[61906]: DEBUG oslo_vmware.api [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333564, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.240236] env[61906]: DEBUG nova.network.neutron [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1004.368703] env[61906]: DEBUG nova.scheduler.client.report [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1004.391131] env[61906]: DEBUG nova.network.neutron [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Updating instance_info_cache with network_info: [{"id": "e0244018-d931-4a96-a25e-228b02a88f56", "address": "fa:16:3e:43:92:23", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0244018-d9", "ovs_interfaceid": "e0244018-d931-4a96-a25e-228b02a88f56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.456493] env[61906]: DEBUG oslo_vmware.api [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333561, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.320754} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.456763] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1004.456954] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1004.457146] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1004.457320] env[61906]: INFO nova.compute.manager [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1004.457556] env[61906]: DEBUG oslo.service.loopingcall [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1004.457750] env[61906]: DEBUG nova.compute.manager [-] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1004.457965] env[61906]: DEBUG nova.network.neutron [-] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1004.465164] env[61906]: DEBUG nova.network.neutron [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Port da64f0e8-39b7-4d88-aa6d-4ac0939a0a20 binding to destination host cpu-1 is already ACTIVE {{(pid=61906) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1004.529255] env[61906]: DEBUG oslo_vmware.api [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Task: {'id': task-1333563, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.27652} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.529418] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1004.529607] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1004.529776] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1004.529981] env[61906]: INFO nova.compute.manager [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1004.530246] env[61906]: DEBUG oslo.service.loopingcall [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1004.530443] env[61906]: DEBUG nova.compute.manager [-] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1004.530585] env[61906]: DEBUG nova.network.neutron [-] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1004.552260] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c5db4317-d69d-4425-8fe1-14aa93f877ed tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "b1159533-c970-49d6-af42-b954b20d92fb" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.563s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.596137] env[61906]: DEBUG nova.compute.manager [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1004.610059] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333556, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.729093] env[61906]: DEBUG nova.compute.manager [req-19d8bdc3-d270-422c-bcd0-b21749d724b7 req-937064d5-7b8c-46d9-bc93-ffbd89a8740a service nova] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Received event network-vif-deleted-5cee4cef-6d3a-4191-836c-a40217b08b3c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1004.729402] env[61906]: INFO nova.compute.manager [req-19d8bdc3-d270-422c-bcd0-b21749d724b7 req-937064d5-7b8c-46d9-bc93-ffbd89a8740a service nova] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Neutron deleted interface 5cee4cef-6d3a-4191-836c-a40217b08b3c; detaching it from the instance and deleting it from the info cache [ 1004.729636] env[61906]: DEBUG nova.network.neutron [req-19d8bdc3-d270-422c-bcd0-b21749d724b7 req-937064d5-7b8c-46d9-bc93-ffbd89a8740a service nova] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.737779] env[61906]: DEBUG oslo_vmware.api [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333564, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.874038] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1004.874038] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.317s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.874038] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.707s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.875140] env[61906]: DEBUG nova.objects.instance [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lazy-loading 'resources' on Instance uuid 150bf47d-7da9-4e13-ad00-7a8a25b9504e {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1004.893401] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Releasing lock "refresh_cache-b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.893718] env[61906]: DEBUG nova.compute.manager [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Instance network_info: |[{"id": "e0244018-d931-4a96-a25e-228b02a88f56", "address": "fa:16:3e:43:92:23", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0244018-d9", "ovs_interfaceid": "e0244018-d931-4a96-a25e-228b02a88f56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1004.894237] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:43:92:23', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35ac9709-fd8b-4630-897a-68ed629d1b11', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e0244018-d931-4a96-a25e-228b02a88f56', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1004.901931] env[61906]: DEBUG oslo.service.loopingcall [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1004.902483] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1004.902688] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-44eea89d-036d-40ef-864f-5545441f7aba {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.924585] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1004.924585] env[61906]: value = "task-1333565" [ 1004.924585] env[61906]: _type = "Task" [ 1004.924585] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.935314] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333565, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.112794] env[61906]: DEBUG oslo_vmware.api [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333556, 'name': PowerOnVM_Task, 'duration_secs': 2.041707} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.113154] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1005.113337] env[61906]: INFO nova.compute.manager [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Took 17.83 seconds to spawn the instance on the hypervisor. [ 1005.113518] env[61906]: DEBUG nova.compute.manager [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1005.114317] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d288a87-8623-4b95-985d-b1deabef5a5b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.124801] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.126261] env[61906]: DEBUG nova.compute.manager [req-dea6a063-95c8-4d1c-a5a4-dd2ebbf2fd60 req-f56fd57c-053a-4cba-b2b9-96520af6f236 service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Received event network-changed-e0244018-d931-4a96-a25e-228b02a88f56 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1005.126442] env[61906]: DEBUG nova.compute.manager [req-dea6a063-95c8-4d1c-a5a4-dd2ebbf2fd60 req-f56fd57c-053a-4cba-b2b9-96520af6f236 service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Refreshing instance network info cache due to event network-changed-e0244018-d931-4a96-a25e-228b02a88f56. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1005.126650] env[61906]: DEBUG oslo_concurrency.lockutils [req-dea6a063-95c8-4d1c-a5a4-dd2ebbf2fd60 req-f56fd57c-053a-4cba-b2b9-96520af6f236 service nova] Acquiring lock "refresh_cache-b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.126787] env[61906]: DEBUG oslo_concurrency.lockutils [req-dea6a063-95c8-4d1c-a5a4-dd2ebbf2fd60 req-f56fd57c-053a-4cba-b2b9-96520af6f236 service nova] Acquired lock "refresh_cache-b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.126948] env[61906]: DEBUG nova.network.neutron [req-dea6a063-95c8-4d1c-a5a4-dd2ebbf2fd60 req-f56fd57c-053a-4cba-b2b9-96520af6f236 service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Refreshing network info cache for port e0244018-d931-4a96-a25e-228b02a88f56 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1005.192413] env[61906]: DEBUG nova.network.neutron [-] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.234068] env[61906]: DEBUG oslo_vmware.api [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333564, 'name': ReconfigVM_Task, 'duration_secs': 0.601778} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.234068] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Reconfigured VM instance instance-00000059 to attach disk [datastore1] a3e3ad00-6921-4072-8cb1-d80302883513/a3e3ad00-6921-4072-8cb1-d80302883513.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1005.234068] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Updating instance 'a3e3ad00-6921-4072-8cb1-d80302883513' progress to 50 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1005.238873] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-83722691-2b74-406d-9953-40b631e036a7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.249066] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36d6dda0-5813-4b63-9970-073d9b84937f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.287807] env[61906]: DEBUG nova.compute.manager [req-19d8bdc3-d270-422c-bcd0-b21749d724b7 req-937064d5-7b8c-46d9-bc93-ffbd89a8740a service nova] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Detach interface failed, port_id=5cee4cef-6d3a-4191-836c-a40217b08b3c, reason: Instance 34d22e43-882f-45c2-8259-e77424ce0fb6 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1005.367463] env[61906]: DEBUG nova.network.neutron [-] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.435283] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333565, 'name': CreateVM_Task, 'duration_secs': 0.4522} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.435457] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1005.436133] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.436310] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.436626] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1005.436877] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe44d29e-01c8-4aed-ab63-0c8fd5067bb9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.442626] env[61906]: DEBUG oslo_vmware.api [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 1005.442626] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52c6b85b-1617-fff6-27ff-4a9a2fcf1028" [ 1005.442626] env[61906]: _type = "Task" [ 1005.442626] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.452386] env[61906]: DEBUG oslo_vmware.api [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52c6b85b-1617-fff6-27ff-4a9a2fcf1028, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.487885] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "458a1669-a62b-4313-874a-e49809d5c034-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.488139] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "458a1669-a62b-4313-874a-e49809d5c034-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.488319] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "458a1669-a62b-4313-874a-e49809d5c034-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.566169] env[61906]: DEBUG nova.objects.instance [None req-fb892012-3b42-4a2b-826d-7880f928247d tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lazy-loading 'flavor' on Instance uuid b1159533-c970-49d6-af42-b954b20d92fb {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.568681] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25c63459-5798-4603-9f41-c6da8e325c14 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.577241] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25ad2a8a-c4a2-4858-9514-3da56958ac63 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.607978] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d707ece3-d71d-4187-8d29-df6b0946cdc1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.615656] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f8a03a9-6752-45c0-b5c8-a1cff031d51c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.634206] env[61906]: DEBUG nova.compute.provider_tree [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1005.637965] env[61906]: INFO nova.compute.manager [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Took 22.72 seconds to build instance. [ 1005.694271] env[61906]: INFO nova.compute.manager [-] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Took 1.24 seconds to deallocate network for instance. [ 1005.740298] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb37c18-8232-40ff-98f3-7a9459e23502 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.769554] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f7551c6-519b-49a1-93d3-757a6182bb66 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.790445] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Updating instance 'a3e3ad00-6921-4072-8cb1-d80302883513' progress to 67 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1005.870400] env[61906]: INFO nova.compute.manager [-] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Took 1.34 seconds to deallocate network for instance. [ 1005.941834] env[61906]: DEBUG nova.network.neutron [req-dea6a063-95c8-4d1c-a5a4-dd2ebbf2fd60 req-f56fd57c-053a-4cba-b2b9-96520af6f236 service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Updated VIF entry in instance network info cache for port e0244018-d931-4a96-a25e-228b02a88f56. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1005.941834] env[61906]: DEBUG nova.network.neutron [req-dea6a063-95c8-4d1c-a5a4-dd2ebbf2fd60 req-f56fd57c-053a-4cba-b2b9-96520af6f236 service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Updating instance_info_cache with network_info: [{"id": "e0244018-d931-4a96-a25e-228b02a88f56", "address": "fa:16:3e:43:92:23", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0244018-d9", "ovs_interfaceid": "e0244018-d931-4a96-a25e-228b02a88f56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.953830] env[61906]: DEBUG oslo_vmware.api [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52c6b85b-1617-fff6-27ff-4a9a2fcf1028, 'name': SearchDatastore_Task, 'duration_secs': 0.012908} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.954161] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.954409] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1005.954698] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.954787] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.954974] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1005.955798] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-12a7140d-7f75-40ff-8212-76b63dae6d1c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.966102] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1005.966316] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1005.967063] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a084769-92c8-4a5d-93d2-3297f28f4784 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.973885] env[61906]: DEBUG oslo_vmware.api [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 1005.973885] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]525995e7-504c-5e6a-f20d-487e944a1a82" [ 1005.973885] env[61906]: _type = "Task" [ 1005.973885] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.988064] env[61906]: DEBUG oslo_vmware.api [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]525995e7-504c-5e6a-f20d-487e944a1a82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.074714] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fb892012-3b42-4a2b-826d-7880f928247d tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "refresh_cache-b1159533-c970-49d6-af42-b954b20d92fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.075743] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fb892012-3b42-4a2b-826d-7880f928247d tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquired lock "refresh_cache-b1159533-c970-49d6-af42-b954b20d92fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.075743] env[61906]: DEBUG nova.network.neutron [None req-fb892012-3b42-4a2b-826d-7880f928247d tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1006.075743] env[61906]: DEBUG nova.objects.instance [None req-fb892012-3b42-4a2b-826d-7880f928247d tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lazy-loading 'info_cache' on Instance uuid b1159533-c970-49d6-af42-b954b20d92fb {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1006.137637] env[61906]: DEBUG nova.scheduler.client.report [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1006.140905] env[61906]: DEBUG oslo_concurrency.lockutils [None req-76235886-2262-460c-a1d5-ee383a4b9519 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "37124c5c-2021-415b-925b-52e23a3d2973" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.276s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.210714] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.330160] env[61906]: DEBUG nova.network.neutron [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Port e3ce81bf-b27e-47ea-9c48-a17e608a9cd6 binding to destination host cpu-1 is already ACTIVE {{(pid=61906) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1006.380505] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.443816] env[61906]: DEBUG oslo_concurrency.lockutils [req-dea6a063-95c8-4d1c-a5a4-dd2ebbf2fd60 req-f56fd57c-053a-4cba-b2b9-96520af6f236 service nova] Releasing lock "refresh_cache-b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.444155] env[61906]: DEBUG nova.compute.manager [req-dea6a063-95c8-4d1c-a5a4-dd2ebbf2fd60 req-f56fd57c-053a-4cba-b2b9-96520af6f236 service nova] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Received event network-vif-deleted-d21f5215-2510-4ea0-922d-654e42e6f1d1 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1006.444403] env[61906]: INFO nova.compute.manager [req-dea6a063-95c8-4d1c-a5a4-dd2ebbf2fd60 req-f56fd57c-053a-4cba-b2b9-96520af6f236 service nova] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Neutron deleted interface d21f5215-2510-4ea0-922d-654e42e6f1d1; detaching it from the instance and deleting it from the info cache [ 1006.444618] env[61906]: DEBUG nova.network.neutron [req-dea6a063-95c8-4d1c-a5a4-dd2ebbf2fd60 req-f56fd57c-053a-4cba-b2b9-96520af6f236 service nova] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.485776] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "37124c5c-2021-415b-925b-52e23a3d2973" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.485995] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "37124c5c-2021-415b-925b-52e23a3d2973" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.486230] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "37124c5c-2021-415b-925b-52e23a3d2973-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.486417] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "37124c5c-2021-415b-925b-52e23a3d2973-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.486586] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "37124c5c-2021-415b-925b-52e23a3d2973-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.488445] env[61906]: DEBUG oslo_vmware.api [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]525995e7-504c-5e6a-f20d-487e944a1a82, 'name': SearchDatastore_Task, 'duration_secs': 0.015133} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.488895] env[61906]: INFO nova.compute.manager [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Terminating instance [ 1006.491080] env[61906]: DEBUG nova.compute.manager [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1006.491280] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1006.491517] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e89532a-34c0-4b16-bdd3-ad3e12f3caf3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.496275] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7d37edf-3f42-491a-b12c-d45313f6ef04 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.505479] env[61906]: DEBUG oslo_vmware.api [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 1006.505479] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]527633a0-d6ca-5813-a44f-c0db8f739f48" [ 1006.505479] env[61906]: _type = "Task" [ 1006.505479] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.507694] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1006.510833] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-29cd45ef-0d34-4faa-a0e5-47c1f7b779d4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.518236] env[61906]: DEBUG oslo_vmware.api [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]527633a0-d6ca-5813-a44f-c0db8f739f48, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.523034] env[61906]: DEBUG oslo_vmware.api [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 1006.523034] env[61906]: value = "task-1333566" [ 1006.523034] env[61906]: _type = "Task" [ 1006.523034] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.532746] env[61906]: DEBUG oslo_vmware.api [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333566, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.549809] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "refresh_cache-458a1669-a62b-4313-874a-e49809d5c034" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.550017] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired lock "refresh_cache-458a1669-a62b-4313-874a-e49809d5c034" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.550222] env[61906]: DEBUG nova.network.neutron [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1006.579076] env[61906]: DEBUG nova.objects.base [None req-fb892012-3b42-4a2b-826d-7880f928247d tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1006.643121] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.770s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.646202] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.521s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.648607] env[61906]: INFO nova.compute.claims [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1006.668731] env[61906]: INFO nova.scheduler.client.report [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Deleted allocations for instance 150bf47d-7da9-4e13-ad00-7a8a25b9504e [ 1006.947452] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-711a8065-2eee-4e69-ab19-54892a42b191 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.959475] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da99ff11-ef40-46cc-8a64-bc0f1d7bca54 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.988313] env[61906]: DEBUG nova.compute.manager [req-dea6a063-95c8-4d1c-a5a4-dd2ebbf2fd60 req-f56fd57c-053a-4cba-b2b9-96520af6f236 service nova] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Detach interface failed, port_id=d21f5215-2510-4ea0-922d-654e42e6f1d1, reason: Instance cec87a1e-4176-45bb-a0c7-a594cb399170 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1007.016826] env[61906]: DEBUG oslo_vmware.api [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]527633a0-d6ca-5813-a44f-c0db8f739f48, 'name': SearchDatastore_Task, 'duration_secs': 0.01439} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.017098] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.017363] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] b3e9ab00-8627-4a57-839e-68be5f794b28/b3e9ab00-8627-4a57-839e-68be5f794b28.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1007.017616] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3501e8c4-cb5c-40c4-a5b3-584a717cd1c1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.024616] env[61906]: DEBUG oslo_vmware.api [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 1007.024616] env[61906]: value = "task-1333567" [ 1007.024616] env[61906]: _type = "Task" [ 1007.024616] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.035250] env[61906]: DEBUG oslo_vmware.api [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333566, 'name': PowerOffVM_Task, 'duration_secs': 0.170573} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.038056] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1007.038333] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1007.038618] env[61906]: DEBUG oslo_vmware.api [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333567, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.038832] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d00aace2-eed5-4ede-961b-163dc5c1fcff {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.104651] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1007.105027] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1007.105252] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Deleting the datastore file [datastore2] 37124c5c-2021-415b-925b-52e23a3d2973 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1007.105560] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0762a1eb-baa9-46b1-99db-828a3edf88ac {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.118346] env[61906]: DEBUG oslo_vmware.api [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 1007.118346] env[61906]: value = "task-1333569" [ 1007.118346] env[61906]: _type = "Task" [ 1007.118346] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.126966] env[61906]: DEBUG oslo_vmware.api [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333569, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.176660] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0ade862-22be-4555-8d47-86a4d8aaf602 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "150bf47d-7da9-4e13-ad00-7a8a25b9504e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.306s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.357029] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "a3e3ad00-6921-4072-8cb1-d80302883513-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.357029] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "a3e3ad00-6921-4072-8cb1-d80302883513-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.357029] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "a3e3ad00-6921-4072-8cb1-d80302883513-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.431794] env[61906]: DEBUG nova.network.neutron [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Updating instance_info_cache with network_info: [{"id": "da64f0e8-39b7-4d88-aa6d-4ac0939a0a20", "address": "fa:16:3e:3b:f8:9b", "network": {"id": "af039bdd-ffd2-483d-9655-e7df48936290", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-111531540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f7fc2042e6549cabe98838990f35842", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda64f0e8-39", "ovs_interfaceid": "da64f0e8-39b7-4d88-aa6d-4ac0939a0a20", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.486263] env[61906]: DEBUG nova.network.neutron [None req-fb892012-3b42-4a2b-826d-7880f928247d tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Updating instance_info_cache with network_info: [{"id": "ba5a12df-7917-4b2d-ac9e-7ef1380dabf7", "address": "fa:16:3e:d3:84:8a", "network": {"id": "1c55d0cc-4c30-4b89-8736-3e96c5c9c507", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-245940583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fbdcd2dc0d04e83b9f82a19bdfd0023", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bcd9d2d-25c8-41ad-9a4a-93b9029ba993", "external-id": "nsx-vlan-transportzone-734", "segmentation_id": 734, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba5a12df-79", "ovs_interfaceid": "ba5a12df-7917-4b2d-ac9e-7ef1380dabf7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.538194] env[61906]: DEBUG oslo_vmware.api [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333567, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.452789} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.538414] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] b3e9ab00-8627-4a57-839e-68be5f794b28/b3e9ab00-8627-4a57-839e-68be5f794b28.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1007.538704] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1007.538888] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8e0cdc24-bf85-4222-ad1d-3c1f346edeb0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.548079] env[61906]: DEBUG oslo_vmware.api [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 1007.548079] env[61906]: value = "task-1333570" [ 1007.548079] env[61906]: _type = "Task" [ 1007.548079] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.557811] env[61906]: DEBUG oslo_vmware.api [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333570, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.630281] env[61906]: DEBUG oslo_vmware.api [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333569, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.211866} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.630634] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1007.630833] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1007.631120] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1007.631335] env[61906]: INFO nova.compute.manager [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1007.631592] env[61906]: DEBUG oslo.service.loopingcall [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1007.631793] env[61906]: DEBUG nova.compute.manager [-] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1007.632057] env[61906]: DEBUG nova.network.neutron [-] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1007.844873] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b55142dc-8d1d-4b5b-8b89-80728ba651ef {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.854147] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f632148-65ad-456b-a70c-b14f0f95f7bb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.890384] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05670133-0c8b-4800-9490-42b374310989 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.898697] env[61906]: DEBUG nova.compute.manager [req-225cf381-128d-4d14-9238-4a631d939a39 req-ed925b39-db55-434a-9c13-b70d0725874d service nova] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Received event network-vif-deleted-077e6829-d958-4859-903a-8cad7602c935 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1007.898880] env[61906]: INFO nova.compute.manager [req-225cf381-128d-4d14-9238-4a631d939a39 req-ed925b39-db55-434a-9c13-b70d0725874d service nova] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Neutron deleted interface 077e6829-d958-4859-903a-8cad7602c935; detaching it from the instance and deleting it from the info cache [ 1007.899123] env[61906]: DEBUG nova.network.neutron [req-225cf381-128d-4d14-9238-4a631d939a39 req-ed925b39-db55-434a-9c13-b70d0725874d service nova] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.904712] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f635e486-a20a-4752-a542-0e37cbaaa41a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.920200] env[61906]: DEBUG nova.compute.provider_tree [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1007.934337] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Releasing lock "refresh_cache-458a1669-a62b-4313-874a-e49809d5c034" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.988526] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fb892012-3b42-4a2b-826d-7880f928247d tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Releasing lock "refresh_cache-b1159533-c970-49d6-af42-b954b20d92fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.058038] env[61906]: DEBUG oslo_vmware.api [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333570, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.114175} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.058272] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1008.059049] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68ad3c7d-a969-4d69-abaa-e3351031e1fd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.081694] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] b3e9ab00-8627-4a57-839e-68be5f794b28/b3e9ab00-8627-4a57-839e-68be5f794b28.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1008.081964] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f1d1f094-211c-413a-9042-19a10acbd1bd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.102134] env[61906]: DEBUG oslo_vmware.api [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 1008.102134] env[61906]: value = "task-1333571" [ 1008.102134] env[61906]: _type = "Task" [ 1008.102134] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.110054] env[61906]: DEBUG oslo_vmware.api [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333571, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.261040] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cff48d68-48d2-4ed7-b45c-aecea6a39e32 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "7504dfa6-bedc-4701-b4fc-60e19e742276" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.261343] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cff48d68-48d2-4ed7-b45c-aecea6a39e32 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "7504dfa6-bedc-4701-b4fc-60e19e742276" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.397894] env[61906]: DEBUG nova.network.neutron [-] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.404961] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-82f0d392-7614-4004-b954-08467e21fd03 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.414292] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60fa3690-b8d1-4156-8d45-5508aaf555bb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.426621] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "refresh_cache-a3e3ad00-6921-4072-8cb1-d80302883513" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.426816] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquired lock "refresh_cache-a3e3ad00-6921-4072-8cb1-d80302883513" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.426999] env[61906]: DEBUG nova.network.neutron [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1008.429135] env[61906]: DEBUG nova.scheduler.client.report [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1008.454109] env[61906]: DEBUG nova.compute.manager [req-225cf381-128d-4d14-9238-4a631d939a39 req-ed925b39-db55-434a-9c13-b70d0725874d service nova] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Detach interface failed, port_id=077e6829-d958-4859-903a-8cad7602c935, reason: Instance 37124c5c-2021-415b-925b-52e23a3d2973 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1008.455855] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f63469a-456d-4118-9243-93457a3aab97 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.475640] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-970c8926-f803-425e-9778-19a3a745c206 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.485762] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Updating instance '458a1669-a62b-4313-874a-e49809d5c034' progress to 83 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1008.495748] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb892012-3b42-4a2b-826d-7880f928247d tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1008.495748] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1c5976ae-b347-409c-aaad-6fa467c5497a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.502551] env[61906]: DEBUG oslo_vmware.api [None req-fb892012-3b42-4a2b-826d-7880f928247d tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 1008.502551] env[61906]: value = "task-1333572" [ 1008.502551] env[61906]: _type = "Task" [ 1008.502551] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.511040] env[61906]: DEBUG oslo_vmware.api [None req-fb892012-3b42-4a2b-826d-7880f928247d tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333572, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.612787] env[61906]: DEBUG oslo_vmware.api [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333571, 'name': ReconfigVM_Task, 'duration_secs': 0.278876} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.613342] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Reconfigured VM instance instance-0000005e to attach disk [datastore1] b3e9ab00-8627-4a57-839e-68be5f794b28/b3e9ab00-8627-4a57-839e-68be5f794b28.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1008.614314] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-23591961-f7c6-4508-b09e-ae04709624de {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.623133] env[61906]: DEBUG oslo_vmware.api [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 1008.623133] env[61906]: value = "task-1333573" [ 1008.623133] env[61906]: _type = "Task" [ 1008.623133] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.635616] env[61906]: DEBUG oslo_vmware.api [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333573, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.765989] env[61906]: INFO nova.compute.manager [None req-cff48d68-48d2-4ed7-b45c-aecea6a39e32 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Detaching volume a8641678-cc5c-458c-a0a6-daf5317bf7d7 [ 1008.804105] env[61906]: INFO nova.virt.block_device [None req-cff48d68-48d2-4ed7-b45c-aecea6a39e32 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Attempting to driver detach volume a8641678-cc5c-458c-a0a6-daf5317bf7d7 from mountpoint /dev/sdb [ 1008.804395] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-cff48d68-48d2-4ed7-b45c-aecea6a39e32 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Volume detach. Driver type: vmdk {{(pid=61906) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1008.805031] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-cff48d68-48d2-4ed7-b45c-aecea6a39e32 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284799', 'volume_id': 'a8641678-cc5c-458c-a0a6-daf5317bf7d7', 'name': 'volume-a8641678-cc5c-458c-a0a6-daf5317bf7d7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7504dfa6-bedc-4701-b4fc-60e19e742276', 'attached_at': '', 'detached_at': '', 'volume_id': 'a8641678-cc5c-458c-a0a6-daf5317bf7d7', 'serial': 'a8641678-cc5c-458c-a0a6-daf5317bf7d7'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1008.805698] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5bb0ef6-49af-4cbd-88ad-50265a742064 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.829904] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37ec4d52-fc96-4d7f-9a8b-b6e08cc1be77 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.837909] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a50a0f7f-a9a4-40c9-a02d-5958d1cbcd9e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.858531] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c975620-d04a-40f8-983c-acc7f7f0383b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.877303] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-cff48d68-48d2-4ed7-b45c-aecea6a39e32 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] The volume has not been displaced from its original location: [datastore2] volume-a8641678-cc5c-458c-a0a6-daf5317bf7d7/volume-a8641678-cc5c-458c-a0a6-daf5317bf7d7.vmdk. No consolidation needed. {{(pid=61906) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1008.882789] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-cff48d68-48d2-4ed7-b45c-aecea6a39e32 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Reconfiguring VM instance instance-00000046 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1008.883205] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0126d53c-3c83-4768-81a0-2a3ee35b665d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.900688] env[61906]: INFO nova.compute.manager [-] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Took 1.27 seconds to deallocate network for instance. [ 1008.904322] env[61906]: DEBUG oslo_vmware.api [None req-cff48d68-48d2-4ed7-b45c-aecea6a39e32 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1008.904322] env[61906]: value = "task-1333574" [ 1008.904322] env[61906]: _type = "Task" [ 1008.904322] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.918639] env[61906]: DEBUG oslo_vmware.api [None req-cff48d68-48d2-4ed7-b45c-aecea6a39e32 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333574, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.935353] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.289s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.935873] env[61906]: DEBUG nova.compute.manager [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1008.938640] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.728s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.938935] env[61906]: DEBUG nova.objects.instance [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lazy-loading 'resources' on Instance uuid 34d22e43-882f-45c2-8259-e77424ce0fb6 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1008.991369] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1008.992037] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c7ed5de5-cfe1-4f7e-bd80-a03d6bc36e0b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.000351] env[61906]: DEBUG oslo_vmware.api [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 1009.000351] env[61906]: value = "task-1333575" [ 1009.000351] env[61906]: _type = "Task" [ 1009.000351] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.013915] env[61906]: DEBUG oslo_vmware.api [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333575, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.018390] env[61906]: DEBUG oslo_vmware.api [None req-fb892012-3b42-4a2b-826d-7880f928247d tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333572, 'name': PowerOnVM_Task, 'duration_secs': 0.434504} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.018756] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb892012-3b42-4a2b-826d-7880f928247d tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1009.019012] env[61906]: DEBUG nova.compute.manager [None req-fb892012-3b42-4a2b-826d-7880f928247d tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1009.019964] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49c8e596-a11a-436a-8ddf-c26de961e9a1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.133128] env[61906]: DEBUG oslo_vmware.api [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333573, 'name': Rename_Task, 'duration_secs': 0.147958} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.133431] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1009.133688] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d5d090fd-da63-4cd6-911f-5a9a1db8dd51 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.144882] env[61906]: DEBUG oslo_vmware.api [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 1009.144882] env[61906]: value = "task-1333576" [ 1009.144882] env[61906]: _type = "Task" [ 1009.144882] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.152697] env[61906]: DEBUG oslo_vmware.api [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333576, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.162641] env[61906]: DEBUG nova.network.neutron [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Updating instance_info_cache with network_info: [{"id": "e3ce81bf-b27e-47ea-9c48-a17e608a9cd6", "address": "fa:16:3e:03:ef:06", "network": {"id": "ddc08dc9-d08f-4fb9-9bee-1a030ed69f76", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1124067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b33dea0794a48f78b9f519cb269a8c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f1b07b1-e4e5-4842-9090-07fb2c3e124b", "external-id": "nsx-vlan-transportzone-646", "segmentation_id": 646, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3ce81bf-b2", "ovs_interfaceid": "e3ce81bf-b27e-47ea-9c48-a17e608a9cd6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.412156] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.418586] env[61906]: DEBUG oslo_vmware.api [None req-cff48d68-48d2-4ed7-b45c-aecea6a39e32 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333574, 'name': ReconfigVM_Task, 'duration_secs': 0.28498} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.418881] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-cff48d68-48d2-4ed7-b45c-aecea6a39e32 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Reconfigured VM instance instance-00000046 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1009.424034] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e7b90d3-65a4-4141-909f-88ab5138620e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.440375] env[61906]: DEBUG oslo_vmware.api [None req-cff48d68-48d2-4ed7-b45c-aecea6a39e32 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1009.440375] env[61906]: value = "task-1333577" [ 1009.440375] env[61906]: _type = "Task" [ 1009.440375] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.441783] env[61906]: DEBUG nova.compute.utils [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1009.448258] env[61906]: DEBUG nova.compute.manager [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1009.448421] env[61906]: DEBUG nova.network.neutron [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1009.456442] env[61906]: DEBUG oslo_vmware.api [None req-cff48d68-48d2-4ed7-b45c-aecea6a39e32 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333577, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.512475] env[61906]: DEBUG oslo_vmware.api [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333575, 'name': PowerOnVM_Task, 'duration_secs': 0.403614} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.515307] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1009.515515] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e95fd4c9-dfc6-4c3f-b700-e19a39b46191 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Updating instance '458a1669-a62b-4313-874a-e49809d5c034' progress to 100 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1009.520896] env[61906]: DEBUG nova.policy [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b49e5a6aeb1d43e290a645e14861b889', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '53644a4143a24759a9ff2b5e28b84fb4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1009.659331] env[61906]: DEBUG oslo_vmware.api [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333576, 'name': PowerOnVM_Task, 'duration_secs': 0.479382} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.659642] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1009.659857] env[61906]: INFO nova.compute.manager [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Took 7.02 seconds to spawn the instance on the hypervisor. [ 1009.660076] env[61906]: DEBUG nova.compute.manager [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1009.660922] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f544d505-4f5c-4d3a-a1eb-a2ef076c969a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.666524] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Releasing lock "refresh_cache-a3e3ad00-6921-4072-8cb1-d80302883513" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.682168] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e4144e0-d252-48f0-8a2b-1508bb340d06 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.690866] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6aa6776-389d-40e5-a406-8f28b4b5a7a7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.727662] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37c19f23-6086-459d-8bb8-31a46cc040f7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.737754] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af9c1833-2cdb-4725-81e3-6cfd0572cbfd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.756751] env[61906]: DEBUG nova.compute.provider_tree [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1009.845859] env[61906]: DEBUG nova.network.neutron [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Successfully created port: 99bfd128-61db-4547-8a4f-2f3d676352f1 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1009.953633] env[61906]: DEBUG nova.compute.manager [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1009.956374] env[61906]: DEBUG oslo_vmware.api [None req-cff48d68-48d2-4ed7-b45c-aecea6a39e32 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333577, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.187082] env[61906]: INFO nova.compute.manager [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Took 16.71 seconds to build instance. [ 1010.192346] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8fdcb85-737b-473e-9834-27ff71eba607 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.214284] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49e58b44-5be0-4dc4-a388-3643c091cd5d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.223582] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Updating instance 'a3e3ad00-6921-4072-8cb1-d80302883513' progress to 83 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1010.263021] env[61906]: DEBUG nova.scheduler.client.report [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1010.452727] env[61906]: DEBUG oslo_vmware.api [None req-cff48d68-48d2-4ed7-b45c-aecea6a39e32 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333577, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.689792] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c207bce1-d515-4171-a7fc-b3a29dff944c tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "b3e9ab00-8627-4a57-839e-68be5f794b28" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.224s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.729799] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1010.730150] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0f791684-ed6f-4363-8bd2-2c9381803a49 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.738893] env[61906]: DEBUG oslo_vmware.api [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 1010.738893] env[61906]: value = "task-1333578" [ 1010.738893] env[61906]: _type = "Task" [ 1010.738893] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.750259] env[61906]: DEBUG oslo_vmware.api [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333578, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.768707] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.830s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.772569] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.392s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.772569] env[61906]: DEBUG nova.objects.instance [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lazy-loading 'resources' on Instance uuid cec87a1e-4176-45bb-a0c7-a594cb399170 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1010.794563] env[61906]: INFO nova.scheduler.client.report [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Deleted allocations for instance 34d22e43-882f-45c2-8259-e77424ce0fb6 [ 1010.953529] env[61906]: DEBUG oslo_vmware.api [None req-cff48d68-48d2-4ed7-b45c-aecea6a39e32 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333577, 'name': ReconfigVM_Task, 'duration_secs': 1.189106} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.953877] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-cff48d68-48d2-4ed7-b45c-aecea6a39e32 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284799', 'volume_id': 'a8641678-cc5c-458c-a0a6-daf5317bf7d7', 'name': 'volume-a8641678-cc5c-458c-a0a6-daf5317bf7d7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7504dfa6-bedc-4701-b4fc-60e19e742276', 'attached_at': '', 'detached_at': '', 'volume_id': 'a8641678-cc5c-458c-a0a6-daf5317bf7d7', 'serial': 'a8641678-cc5c-458c-a0a6-daf5317bf7d7'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1010.962654] env[61906]: DEBUG nova.compute.manager [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1010.994748] env[61906]: DEBUG nova.virt.hardware [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1010.995071] env[61906]: DEBUG nova.virt.hardware [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1010.995281] env[61906]: DEBUG nova.virt.hardware [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1010.995506] env[61906]: DEBUG nova.virt.hardware [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1010.995690] env[61906]: DEBUG nova.virt.hardware [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1010.995868] env[61906]: DEBUG nova.virt.hardware [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1010.996138] env[61906]: DEBUG nova.virt.hardware [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1010.996397] env[61906]: DEBUG nova.virt.hardware [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1010.996539] env[61906]: DEBUG nova.virt.hardware [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1010.996738] env[61906]: DEBUG nova.virt.hardware [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1010.996956] env[61906]: DEBUG nova.virt.hardware [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1010.997910] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-640a87f2-82e4-4716-a9ef-e6fb4794d8f5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.009153] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7750742-ae6b-4b49-8848-dfcfa0bb1f5f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.249600] env[61906]: DEBUG oslo_vmware.api [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333578, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.301747] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f9d40b8-0da2-49ce-b8b1-8c7585179af3 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "34d22e43-882f-45c2-8259-e77424ce0fb6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.984s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.404989] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7fce435d-d7e3-40a2-9b3f-9a876bd98f5c tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "1a335019-915b-4065-b22f-d2f0af728d6f" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.405882] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7fce435d-d7e3-40a2-9b3f-9a876bd98f5c tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "1a335019-915b-4065-b22f-d2f0af728d6f" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.405882] env[61906]: INFO nova.compute.manager [None req-7fce435d-d7e3-40a2-9b3f-9a876bd98f5c tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Rebooting instance [ 1011.518618] env[61906]: DEBUG nova.objects.instance [None req-cff48d68-48d2-4ed7-b45c-aecea6a39e32 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lazy-loading 'flavor' on Instance uuid 7504dfa6-bedc-4701-b4fc-60e19e742276 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1011.527544] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e9ba9a1-c822-4472-9970-d265618f75df {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.533730] env[61906]: DEBUG nova.compute.manager [req-78034a88-874a-428c-99c6-ee85d676f844 req-eb124fa1-78eb-43b4-aee5-cf1433143012 service nova] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Received event network-vif-plugged-99bfd128-61db-4547-8a4f-2f3d676352f1 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1011.533828] env[61906]: DEBUG oslo_concurrency.lockutils [req-78034a88-874a-428c-99c6-ee85d676f844 req-eb124fa1-78eb-43b4-aee5-cf1433143012 service nova] Acquiring lock "df605d77-7766-47b3-bd27-97361ff8ae4d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.534123] env[61906]: DEBUG oslo_concurrency.lockutils [req-78034a88-874a-428c-99c6-ee85d676f844 req-eb124fa1-78eb-43b4-aee5-cf1433143012 service nova] Lock "df605d77-7766-47b3-bd27-97361ff8ae4d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.534305] env[61906]: DEBUG oslo_concurrency.lockutils [req-78034a88-874a-428c-99c6-ee85d676f844 req-eb124fa1-78eb-43b4-aee5-cf1433143012 service nova] Lock "df605d77-7766-47b3-bd27-97361ff8ae4d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.534533] env[61906]: DEBUG nova.compute.manager [req-78034a88-874a-428c-99c6-ee85d676f844 req-eb124fa1-78eb-43b4-aee5-cf1433143012 service nova] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] No waiting events found dispatching network-vif-plugged-99bfd128-61db-4547-8a4f-2f3d676352f1 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1011.534731] env[61906]: WARNING nova.compute.manager [req-78034a88-874a-428c-99c6-ee85d676f844 req-eb124fa1-78eb-43b4-aee5-cf1433143012 service nova] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Received unexpected event network-vif-plugged-99bfd128-61db-4547-8a4f-2f3d676352f1 for instance with vm_state building and task_state spawning. [ 1011.539123] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07e8fd9c-88f2-42fd-a4a9-66c259204c7d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.574565] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7b43b60-7d3d-4888-afe9-1ca7e043bfb1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.584674] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca5a43bb-ad47-4e86-8220-886cdd904859 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.605604] env[61906]: DEBUG nova.compute.provider_tree [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.750427] env[61906]: DEBUG oslo_vmware.api [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333578, 'name': PowerOnVM_Task, 'duration_secs': 0.751641} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.750707] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1011.750914] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aac04a2f-a8a8-4341-bd6d-e7ff3b1d9122 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Updating instance 'a3e3ad00-6921-4072-8cb1-d80302883513' progress to 100 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1011.760565] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "458a1669-a62b-4313-874a-e49809d5c034" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.760886] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "458a1669-a62b-4313-874a-e49809d5c034" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.761106] env[61906]: DEBUG nova.compute.manager [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Going to confirm migration 2 {{(pid=61906) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1011.765133] env[61906]: DEBUG nova.network.neutron [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Successfully updated port: 99bfd128-61db-4547-8a4f-2f3d676352f1 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1011.934484] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7fce435d-d7e3-40a2-9b3f-9a876bd98f5c tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.934677] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7fce435d-d7e3-40a2-9b3f-9a876bd98f5c tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.934854] env[61906]: DEBUG nova.network.neutron [None req-7fce435d-d7e3-40a2-9b3f-9a876bd98f5c tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1012.111691] env[61906]: DEBUG nova.scheduler.client.report [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1012.268779] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "refresh_cache-df605d77-7766-47b3-bd27-97361ff8ae4d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.268779] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired lock "refresh_cache-df605d77-7766-47b3-bd27-97361ff8ae4d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.269046] env[61906]: DEBUG nova.network.neutron [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1012.301695] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "refresh_cache-458a1669-a62b-4313-874a-e49809d5c034" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.301891] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired lock "refresh_cache-458a1669-a62b-4313-874a-e49809d5c034" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.302101] env[61906]: DEBUG nova.network.neutron [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1012.302324] env[61906]: DEBUG nova.objects.instance [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lazy-loading 'info_cache' on Instance uuid 458a1669-a62b-4313-874a-e49809d5c034 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1012.524884] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cff48d68-48d2-4ed7-b45c-aecea6a39e32 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "7504dfa6-bedc-4701-b4fc-60e19e742276" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.263s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.617394] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.845s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.620062] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.208s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.620722] env[61906]: DEBUG nova.objects.instance [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lazy-loading 'resources' on Instance uuid 37124c5c-2021-415b-925b-52e23a3d2973 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1012.648355] env[61906]: INFO nova.scheduler.client.report [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Deleted allocations for instance cec87a1e-4176-45bb-a0c7-a594cb399170 [ 1012.733734] env[61906]: DEBUG nova.network.neutron [None req-7fce435d-d7e3-40a2-9b3f-9a876bd98f5c tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Updating instance_info_cache with network_info: [{"id": "a497b774-60ea-485d-a564-ba7b978b560e", "address": "fa:16:3e:0d:cf:6e", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa497b774-60", "ovs_interfaceid": "a497b774-60ea-485d-a564-ba7b978b560e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.831117] env[61906]: DEBUG nova.network.neutron [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1013.094573] env[61906]: DEBUG nova.compute.manager [req-fd0a0398-14ee-4efd-aa54-53e98406d120 req-4f4f6637-7ba8-4072-9779-0e3cd143eedb service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Received event network-changed-e0244018-d931-4a96-a25e-228b02a88f56 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1013.094769] env[61906]: DEBUG nova.compute.manager [req-fd0a0398-14ee-4efd-aa54-53e98406d120 req-4f4f6637-7ba8-4072-9779-0e3cd143eedb service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Refreshing instance network info cache due to event network-changed-e0244018-d931-4a96-a25e-228b02a88f56. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1013.094978] env[61906]: DEBUG oslo_concurrency.lockutils [req-fd0a0398-14ee-4efd-aa54-53e98406d120 req-4f4f6637-7ba8-4072-9779-0e3cd143eedb service nova] Acquiring lock "refresh_cache-b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.096069] env[61906]: DEBUG oslo_concurrency.lockutils [req-fd0a0398-14ee-4efd-aa54-53e98406d120 req-4f4f6637-7ba8-4072-9779-0e3cd143eedb service nova] Acquired lock "refresh_cache-b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.096069] env[61906]: DEBUG nova.network.neutron [req-fd0a0398-14ee-4efd-aa54-53e98406d120 req-4f4f6637-7ba8-4072-9779-0e3cd143eedb service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Refreshing network info cache for port e0244018-d931-4a96-a25e-228b02a88f56 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1013.099448] env[61906]: DEBUG nova.network.neutron [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Updating instance_info_cache with network_info: [{"id": "99bfd128-61db-4547-8a4f-2f3d676352f1", "address": "fa:16:3e:4a:44:67", "network": {"id": "bc9185e9-f26b-489c-84ee-e6db41ac6283", "bridge": "br-int", "label": "tempest-ServersTestJSON-1167747148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53644a4143a24759a9ff2b5e28b84fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99bfd128-61", "ovs_interfaceid": "99bfd128-61db-4547-8a4f-2f3d676352f1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.158330] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4aa97a8b-9239-4bb8-a5da-252c47c49ee6 tempest-MultipleCreateTestJSON-1809540523 tempest-MultipleCreateTestJSON-1809540523-project-member] Lock "cec87a1e-4176-45bb-a0c7-a594cb399170" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.772s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.239027] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7fce435d-d7e3-40a2-9b3f-9a876bd98f5c tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Releasing lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.240969] env[61906]: DEBUG nova.compute.manager [None req-7fce435d-d7e3-40a2-9b3f-9a876bd98f5c tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1013.242680] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3da2220f-e618-46b0-9566-011270d2ba72 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.293907] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e62eba57-54e4-458c-94e0-0fc14d594a3f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.305226] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7340a6fa-22c4-4648-9a20-0c7a22098052 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.340128] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baf30014-a790-4ef8-b2d8-9f81bc0eb7ab {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.348443] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38b07101-9be5-42c1-bd22-5157283411d5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.365217] env[61906]: DEBUG nova.compute.provider_tree [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1013.591605] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "7504dfa6-bedc-4701-b4fc-60e19e742276" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.591858] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "7504dfa6-bedc-4701-b4fc-60e19e742276" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.592093] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "7504dfa6-bedc-4701-b4fc-60e19e742276-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.592288] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "7504dfa6-bedc-4701-b4fc-60e19e742276-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.592463] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "7504dfa6-bedc-4701-b4fc-60e19e742276-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.594714] env[61906]: INFO nova.compute.manager [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Terminating instance [ 1013.596511] env[61906]: DEBUG nova.compute.manager [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1013.596714] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1013.597556] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-665983f6-0dc9-4898-ac12-1fc4565a4a58 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.602738] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Releasing lock "refresh_cache-df605d77-7766-47b3-bd27-97361ff8ae4d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.603040] env[61906]: DEBUG nova.compute.manager [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Instance network_info: |[{"id": "99bfd128-61db-4547-8a4f-2f3d676352f1", "address": "fa:16:3e:4a:44:67", "network": {"id": "bc9185e9-f26b-489c-84ee-e6db41ac6283", "bridge": "br-int", "label": "tempest-ServersTestJSON-1167747148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53644a4143a24759a9ff2b5e28b84fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99bfd128-61", "ovs_interfaceid": "99bfd128-61db-4547-8a4f-2f3d676352f1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1013.603812] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4a:44:67', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '99bfd128-61db-4547-8a4f-2f3d676352f1', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1013.611494] env[61906]: DEBUG oslo.service.loopingcall [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1013.617027] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1013.617027] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1013.617027] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4862c769-7247-4957-a8ce-baba1988173f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.629252] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-67da3282-faac-4150-a0c7-0cae2ae959b8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.636859] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1013.636859] env[61906]: value = "task-1333580" [ 1013.636859] env[61906]: _type = "Task" [ 1013.636859] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.638194] env[61906]: DEBUG oslo_vmware.api [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1013.638194] env[61906]: value = "task-1333579" [ 1013.638194] env[61906]: _type = "Task" [ 1013.638194] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.650637] env[61906]: DEBUG oslo_vmware.api [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333579, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.653772] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333580, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.776289] env[61906]: DEBUG nova.compute.manager [req-5fd47e39-a05f-4510-91c2-5d79a3c503c0 req-52a2a730-1702-4248-b049-a6ea690b6b52 service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Received event network-changed-73d0e617-fd5e-4fa2-81c5-710a3beac1a4 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1013.776517] env[61906]: DEBUG nova.compute.manager [req-5fd47e39-a05f-4510-91c2-5d79a3c503c0 req-52a2a730-1702-4248-b049-a6ea690b6b52 service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Refreshing instance network info cache due to event network-changed-73d0e617-fd5e-4fa2-81c5-710a3beac1a4. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1013.776741] env[61906]: DEBUG oslo_concurrency.lockutils [req-5fd47e39-a05f-4510-91c2-5d79a3c503c0 req-52a2a730-1702-4248-b049-a6ea690b6b52 service nova] Acquiring lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.776890] env[61906]: DEBUG oslo_concurrency.lockutils [req-5fd47e39-a05f-4510-91c2-5d79a3c503c0 req-52a2a730-1702-4248-b049-a6ea690b6b52 service nova] Acquired lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.777069] env[61906]: DEBUG nova.network.neutron [req-5fd47e39-a05f-4510-91c2-5d79a3c503c0 req-52a2a730-1702-4248-b049-a6ea690b6b52 service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Refreshing network info cache for port 73d0e617-fd5e-4fa2-81c5-710a3beac1a4 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1013.872802] env[61906]: DEBUG nova.scheduler.client.report [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1013.926677] env[61906]: DEBUG nova.network.neutron [req-fd0a0398-14ee-4efd-aa54-53e98406d120 req-4f4f6637-7ba8-4072-9779-0e3cd143eedb service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Updated VIF entry in instance network info cache for port e0244018-d931-4a96-a25e-228b02a88f56. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1013.927067] env[61906]: DEBUG nova.network.neutron [req-fd0a0398-14ee-4efd-aa54-53e98406d120 req-4f4f6637-7ba8-4072-9779-0e3cd143eedb service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Updating instance_info_cache with network_info: [{"id": "e0244018-d931-4a96-a25e-228b02a88f56", "address": "fa:16:3e:43:92:23", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0244018-d9", "ovs_interfaceid": "e0244018-d931-4a96-a25e-228b02a88f56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.936700] env[61906]: DEBUG nova.network.neutron [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Updating instance_info_cache with network_info: [{"id": "da64f0e8-39b7-4d88-aa6d-4ac0939a0a20", "address": "fa:16:3e:3b:f8:9b", "network": {"id": "af039bdd-ffd2-483d-9655-e7df48936290", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-111531540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f7fc2042e6549cabe98838990f35842", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda64f0e8-39", "ovs_interfaceid": "da64f0e8-39b7-4d88-aa6d-4ac0939a0a20", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.152008] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333580, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.155242] env[61906]: DEBUG oslo_vmware.api [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333579, 'name': PowerOffVM_Task, 'duration_secs': 0.22853} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.155518] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1014.155719] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1014.155999] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9a591222-7a33-41d5-9395-827ea72d8e82 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.250400] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1014.250735] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1014.250832] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Deleting the datastore file [datastore2] 7504dfa6-bedc-4701-b4fc-60e19e742276 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1014.251116] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ca529dc5-0067-422e-8f09-91cb8ac43f07 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.262153] env[61906]: DEBUG oslo_vmware.api [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1014.262153] env[61906]: value = "task-1333582" [ 1014.262153] env[61906]: _type = "Task" [ 1014.262153] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.263072] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7385a4b-1140-4b8e-800e-6414b9be4bb2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.275794] env[61906]: DEBUG oslo_vmware.api [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333582, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.278213] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7fce435d-d7e3-40a2-9b3f-9a876bd98f5c tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Doing hard reboot of VM {{(pid=61906) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 1014.278462] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-2345670a-ca54-46c8-b401-23d0011bc16d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.292058] env[61906]: DEBUG oslo_vmware.api [None req-7fce435d-d7e3-40a2-9b3f-9a876bd98f5c tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1014.292058] env[61906]: value = "task-1333583" [ 1014.292058] env[61906]: _type = "Task" [ 1014.292058] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.306847] env[61906]: DEBUG oslo_vmware.api [None req-7fce435d-d7e3-40a2-9b3f-9a876bd98f5c tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333583, 'name': ResetVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.378161] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.758s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.407246] env[61906]: INFO nova.scheduler.client.report [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Deleted allocations for instance 37124c5c-2021-415b-925b-52e23a3d2973 [ 1014.430812] env[61906]: DEBUG oslo_concurrency.lockutils [req-fd0a0398-14ee-4efd-aa54-53e98406d120 req-4f4f6637-7ba8-4072-9779-0e3cd143eedb service nova] Releasing lock "refresh_cache-b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.438926] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Releasing lock "refresh_cache-458a1669-a62b-4313-874a-e49809d5c034" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.439609] env[61906]: DEBUG nova.objects.instance [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lazy-loading 'migration_context' on Instance uuid 458a1669-a62b-4313-874a-e49809d5c034 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1014.580441] env[61906]: DEBUG nova.network.neutron [req-5fd47e39-a05f-4510-91c2-5d79a3c503c0 req-52a2a730-1702-4248-b049-a6ea690b6b52 service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Updated VIF entry in instance network info cache for port 73d0e617-fd5e-4fa2-81c5-710a3beac1a4. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1014.580916] env[61906]: DEBUG nova.network.neutron [req-5fd47e39-a05f-4510-91c2-5d79a3c503c0 req-52a2a730-1702-4248-b049-a6ea690b6b52 service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Updating instance_info_cache with network_info: [{"id": "73d0e617-fd5e-4fa2-81c5-710a3beac1a4", "address": "fa:16:3e:66:bf:d9", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73d0e617-fd", "ovs_interfaceid": "73d0e617-fd5e-4fa2-81c5-710a3beac1a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.614850] env[61906]: DEBUG oslo_concurrency.lockutils [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "a3e3ad00-6921-4072-8cb1-d80302883513" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.615244] env[61906]: DEBUG oslo_concurrency.lockutils [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "a3e3ad00-6921-4072-8cb1-d80302883513" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.615501] env[61906]: DEBUG nova.compute.manager [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Going to confirm migration 3 {{(pid=61906) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1014.650301] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333580, 'name': CreateVM_Task, 'duration_secs': 0.542652} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.650499] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1014.651237] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.651409] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.651734] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1014.651994] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68a08e27-82f4-416c-a8eb-1a43a93c128d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.657621] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 1014.657621] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]528c65f3-7011-d4f6-43e6-7802cce17c99" [ 1014.657621] env[61906]: _type = "Task" [ 1014.657621] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.665996] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]528c65f3-7011-d4f6-43e6-7802cce17c99, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.777359] env[61906]: DEBUG oslo_vmware.api [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333582, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.363016} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.777632] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1014.777828] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1014.778014] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1014.778204] env[61906]: INFO nova.compute.manager [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1014.778453] env[61906]: DEBUG oslo.service.loopingcall [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1014.778652] env[61906]: DEBUG nova.compute.manager [-] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1014.778752] env[61906]: DEBUG nova.network.neutron [-] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1014.805136] env[61906]: DEBUG oslo_vmware.api [None req-7fce435d-d7e3-40a2-9b3f-9a876bd98f5c tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333583, 'name': ResetVM_Task, 'duration_secs': 0.091288} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.808343] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7fce435d-d7e3-40a2-9b3f-9a876bd98f5c tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Did hard reboot of VM {{(pid=61906) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 1014.808343] env[61906]: DEBUG nova.compute.manager [None req-7fce435d-d7e3-40a2-9b3f-9a876bd98f5c tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1014.808343] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e50b32a-7b2a-471c-a2e6-d4c7c3f5fe72 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.916202] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fc5a8b8c-4280-4500-8d9a-596c09496a89 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "37124c5c-2021-415b-925b-52e23a3d2973" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.430s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.942779] env[61906]: DEBUG nova.objects.base [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Object Instance<458a1669-a62b-4313-874a-e49809d5c034> lazy-loaded attributes: info_cache,migration_context {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1014.943990] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d378187a-20a2-4ed8-b5e0-809d9d2ce800 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.966990] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f2ba13a-ede8-4861-a292-9ac8717fa11a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.976069] env[61906]: DEBUG oslo_vmware.api [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 1014.976069] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52e40392-32c1-6be7-199c-6438f137e0e4" [ 1014.976069] env[61906]: _type = "Task" [ 1014.976069] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.984144] env[61906]: DEBUG oslo_vmware.api [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52e40392-32c1-6be7-199c-6438f137e0e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.084114] env[61906]: DEBUG oslo_concurrency.lockutils [req-5fd47e39-a05f-4510-91c2-5d79a3c503c0 req-52a2a730-1702-4248-b049-a6ea690b6b52 service nova] Releasing lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.084870] env[61906]: DEBUG nova.compute.manager [req-5fd47e39-a05f-4510-91c2-5d79a3c503c0 req-52a2a730-1702-4248-b049-a6ea690b6b52 service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Received event network-changed-e0244018-d931-4a96-a25e-228b02a88f56 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1015.084870] env[61906]: DEBUG nova.compute.manager [req-5fd47e39-a05f-4510-91c2-5d79a3c503c0 req-52a2a730-1702-4248-b049-a6ea690b6b52 service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Refreshing instance network info cache due to event network-changed-e0244018-d931-4a96-a25e-228b02a88f56. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1015.084870] env[61906]: DEBUG oslo_concurrency.lockutils [req-5fd47e39-a05f-4510-91c2-5d79a3c503c0 req-52a2a730-1702-4248-b049-a6ea690b6b52 service nova] Acquiring lock "refresh_cache-b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.085101] env[61906]: DEBUG oslo_concurrency.lockutils [req-5fd47e39-a05f-4510-91c2-5d79a3c503c0 req-52a2a730-1702-4248-b049-a6ea690b6b52 service nova] Acquired lock "refresh_cache-b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.085227] env[61906]: DEBUG nova.network.neutron [req-5fd47e39-a05f-4510-91c2-5d79a3c503c0 req-52a2a730-1702-4248-b049-a6ea690b6b52 service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Refreshing network info cache for port e0244018-d931-4a96-a25e-228b02a88f56 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1015.169523] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]528c65f3-7011-d4f6-43e6-7802cce17c99, 'name': SearchDatastore_Task, 'duration_secs': 0.033639} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.169900] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.170167] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1015.170428] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.170583] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.170803] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1015.171208] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7669d972-0c9e-4054-bc03-22d8aa90d9e1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.176157] env[61906]: DEBUG nova.compute.manager [req-025df1bd-bbbf-4956-bbfc-afe659c4a5c9 req-c9c0a4d5-590f-4f3a-86ce-f5b051bd3a41 service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Received event network-changed-73d0e617-fd5e-4fa2-81c5-710a3beac1a4 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1015.176460] env[61906]: DEBUG nova.compute.manager [req-025df1bd-bbbf-4956-bbfc-afe659c4a5c9 req-c9c0a4d5-590f-4f3a-86ce-f5b051bd3a41 service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Refreshing instance network info cache due to event network-changed-73d0e617-fd5e-4fa2-81c5-710a3beac1a4. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1015.176737] env[61906]: DEBUG oslo_concurrency.lockutils [req-025df1bd-bbbf-4956-bbfc-afe659c4a5c9 req-c9c0a4d5-590f-4f3a-86ce-f5b051bd3a41 service nova] Acquiring lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.177031] env[61906]: DEBUG oslo_concurrency.lockutils [req-025df1bd-bbbf-4956-bbfc-afe659c4a5c9 req-c9c0a4d5-590f-4f3a-86ce-f5b051bd3a41 service nova] Acquired lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.177313] env[61906]: DEBUG nova.network.neutron [req-025df1bd-bbbf-4956-bbfc-afe659c4a5c9 req-c9c0a4d5-590f-4f3a-86ce-f5b051bd3a41 service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Refreshing network info cache for port 73d0e617-fd5e-4fa2-81c5-710a3beac1a4 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1015.190325] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1015.190432] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1015.191259] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf72c98d-22c0-4dfa-9720-9359e3485112 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.200333] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 1015.200333] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5208c6f1-9459-9537-5ff8-b6a688be9f12" [ 1015.200333] env[61906]: _type = "Task" [ 1015.200333] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.215762] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5208c6f1-9459-9537-5ff8-b6a688be9f12, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.232354] env[61906]: DEBUG oslo_concurrency.lockutils [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "refresh_cache-a3e3ad00-6921-4072-8cb1-d80302883513" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.232551] env[61906]: DEBUG oslo_concurrency.lockutils [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquired lock "refresh_cache-a3e3ad00-6921-4072-8cb1-d80302883513" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.232732] env[61906]: DEBUG nova.network.neutron [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1015.232993] env[61906]: DEBUG nova.objects.instance [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lazy-loading 'info_cache' on Instance uuid a3e3ad00-6921-4072-8cb1-d80302883513 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1015.324213] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7fce435d-d7e3-40a2-9b3f-9a876bd98f5c tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "1a335019-915b-4065-b22f-d2f0af728d6f" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.919s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.486501] env[61906]: DEBUG oslo_vmware.api [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52e40392-32c1-6be7-199c-6438f137e0e4, 'name': SearchDatastore_Task, 'duration_secs': 0.016353} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.486820] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.487071] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.712808] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5208c6f1-9459-9537-5ff8-b6a688be9f12, 'name': SearchDatastore_Task, 'duration_secs': 0.018347} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.714023] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9f2e060-8338-4516-b472-0a57034322e4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.720928] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 1015.720928] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5210d6ae-6f26-d122-fc8e-eaf6105c9676" [ 1015.720928] env[61906]: _type = "Task" [ 1015.720928] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.728756] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5210d6ae-6f26-d122-fc8e-eaf6105c9676, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.769274] env[61906]: DEBUG oslo_concurrency.lockutils [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "03d3fb1b-0e46-4544-b01d-498a2baf3b45" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.769712] env[61906]: DEBUG oslo_concurrency.lockutils [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "03d3fb1b-0e46-4544-b01d-498a2baf3b45" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.770043] env[61906]: DEBUG oslo_concurrency.lockutils [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "03d3fb1b-0e46-4544-b01d-498a2baf3b45-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.770290] env[61906]: DEBUG oslo_concurrency.lockutils [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "03d3fb1b-0e46-4544-b01d-498a2baf3b45-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.770476] env[61906]: DEBUG oslo_concurrency.lockutils [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "03d3fb1b-0e46-4544-b01d-498a2baf3b45-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.772814] env[61906]: INFO nova.compute.manager [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Terminating instance [ 1015.774814] env[61906]: DEBUG nova.compute.manager [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1015.775015] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1015.775845] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1093e2e0-426d-4f06-9425-d2b8f2da062a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.788120] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1015.788404] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-50e69328-4024-43f8-99ac-9244a94c4c0c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.796345] env[61906]: DEBUG oslo_vmware.api [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 1015.796345] env[61906]: value = "task-1333584" [ 1015.796345] env[61906]: _type = "Task" [ 1015.796345] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.806772] env[61906]: DEBUG oslo_vmware.api [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333584, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.817865] env[61906]: DEBUG nova.network.neutron [req-5fd47e39-a05f-4510-91c2-5d79a3c503c0 req-52a2a730-1702-4248-b049-a6ea690b6b52 service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Updated VIF entry in instance network info cache for port e0244018-d931-4a96-a25e-228b02a88f56. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1015.818261] env[61906]: DEBUG nova.network.neutron [req-5fd47e39-a05f-4510-91c2-5d79a3c503c0 req-52a2a730-1702-4248-b049-a6ea690b6b52 service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Updating instance_info_cache with network_info: [{"id": "e0244018-d931-4a96-a25e-228b02a88f56", "address": "fa:16:3e:43:92:23", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0244018-d9", "ovs_interfaceid": "e0244018-d931-4a96-a25e-228b02a88f56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.830477] env[61906]: DEBUG nova.network.neutron [-] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.946771] env[61906]: DEBUG nova.compute.manager [req-4f1c9c77-23a4-4f55-aadc-a53298fe1b6d req-01134a63-559a-4b49-aa8e-b4b662c3d147 service nova] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Received event network-vif-deleted-175a521b-1c70-41f6-963c-79c978013fa1 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1016.036376] env[61906]: DEBUG nova.network.neutron [req-025df1bd-bbbf-4956-bbfc-afe659c4a5c9 req-c9c0a4d5-590f-4f3a-86ce-f5b051bd3a41 service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Updated VIF entry in instance network info cache for port 73d0e617-fd5e-4fa2-81c5-710a3beac1a4. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1016.036773] env[61906]: DEBUG nova.network.neutron [req-025df1bd-bbbf-4956-bbfc-afe659c4a5c9 req-c9c0a4d5-590f-4f3a-86ce-f5b051bd3a41 service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Updating instance_info_cache with network_info: [{"id": "73d0e617-fd5e-4fa2-81c5-710a3beac1a4", "address": "fa:16:3e:66:bf:d9", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73d0e617-fd", "ovs_interfaceid": "73d0e617-fd5e-4fa2-81c5-710a3beac1a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.181193] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1a7a73c-8523-4249-b1ce-aed144874510 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.190174] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4801aa1-cf23-4562-98d6-2ffcc2aedfd1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.226220] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89cddb7f-e5c1-4097-92b0-966e9c0362df {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.235521] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5210d6ae-6f26-d122-fc8e-eaf6105c9676, 'name': SearchDatastore_Task, 'duration_secs': 0.021488} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.237661] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.237939] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] df605d77-7766-47b3-bd27-97361ff8ae4d/df605d77-7766-47b3-bd27-97361ff8ae4d.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1016.238270] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b54d28f2-c9d2-4a06-a7da-a9e01cd172f4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.241150] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8818ce9-c18f-43bf-ba07-9d8ff1b4a0c4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.259612] env[61906]: DEBUG nova.compute.provider_tree [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1016.265816] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 1016.265816] env[61906]: value = "task-1333585" [ 1016.265816] env[61906]: _type = "Task" [ 1016.265816] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.266298] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "0fe121ad-a539-4c04-bb65-b524cb3d91a8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.266526] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "0fe121ad-a539-4c04-bb65-b524cb3d91a8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.278249] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333585, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.309258] env[61906]: DEBUG oslo_vmware.api [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333584, 'name': PowerOffVM_Task, 'duration_secs': 0.313996} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.309662] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1016.309927] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1016.310327] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c4f942b-a949-410b-bdc1-d3d190ac3af1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.321090] env[61906]: DEBUG oslo_concurrency.lockutils [req-5fd47e39-a05f-4510-91c2-5d79a3c503c0 req-52a2a730-1702-4248-b049-a6ea690b6b52 service nova] Releasing lock "refresh_cache-b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.321392] env[61906]: DEBUG nova.compute.manager [req-5fd47e39-a05f-4510-91c2-5d79a3c503c0 req-52a2a730-1702-4248-b049-a6ea690b6b52 service nova] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Received event network-changed-99bfd128-61db-4547-8a4f-2f3d676352f1 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1016.321572] env[61906]: DEBUG nova.compute.manager [req-5fd47e39-a05f-4510-91c2-5d79a3c503c0 req-52a2a730-1702-4248-b049-a6ea690b6b52 service nova] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Refreshing instance network info cache due to event network-changed-99bfd128-61db-4547-8a4f-2f3d676352f1. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1016.321788] env[61906]: DEBUG oslo_concurrency.lockutils [req-5fd47e39-a05f-4510-91c2-5d79a3c503c0 req-52a2a730-1702-4248-b049-a6ea690b6b52 service nova] Acquiring lock "refresh_cache-df605d77-7766-47b3-bd27-97361ff8ae4d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.321931] env[61906]: DEBUG oslo_concurrency.lockutils [req-5fd47e39-a05f-4510-91c2-5d79a3c503c0 req-52a2a730-1702-4248-b049-a6ea690b6b52 service nova] Acquired lock "refresh_cache-df605d77-7766-47b3-bd27-97361ff8ae4d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.322117] env[61906]: DEBUG nova.network.neutron [req-5fd47e39-a05f-4510-91c2-5d79a3c503c0 req-52a2a730-1702-4248-b049-a6ea690b6b52 service nova] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Refreshing network info cache for port 99bfd128-61db-4547-8a4f-2f3d676352f1 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1016.332278] env[61906]: INFO nova.compute.manager [-] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Took 1.55 seconds to deallocate network for instance. [ 1016.378358] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1016.378646] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1016.378832] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Deleting the datastore file [datastore1] 03d3fb1b-0e46-4544-b01d-498a2baf3b45 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1016.379128] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-53cb3f75-7f67-4e8a-b3eb-95b0aa79504d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.388083] env[61906]: DEBUG oslo_vmware.api [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 1016.388083] env[61906]: value = "task-1333587" [ 1016.388083] env[61906]: _type = "Task" [ 1016.388083] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.398958] env[61906]: DEBUG oslo_vmware.api [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333587, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.465818] env[61906]: DEBUG nova.network.neutron [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Updating instance_info_cache with network_info: [{"id": "e3ce81bf-b27e-47ea-9c48-a17e608a9cd6", "address": "fa:16:3e:03:ef:06", "network": {"id": "ddc08dc9-d08f-4fb9-9bee-1a030ed69f76", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1124067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b33dea0794a48f78b9f519cb269a8c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f1b07b1-e4e5-4842-9090-07fb2c3e124b", "external-id": "nsx-vlan-transportzone-646", "segmentation_id": 646, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3ce81bf-b2", "ovs_interfaceid": "e3ce81bf-b27e-47ea-9c48-a17e608a9cd6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.541632] env[61906]: DEBUG oslo_concurrency.lockutils [req-025df1bd-bbbf-4956-bbfc-afe659c4a5c9 req-c9c0a4d5-590f-4f3a-86ce-f5b051bd3a41 service nova] Releasing lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.774539] env[61906]: DEBUG nova.compute.manager [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1016.783591] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333585, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.789554] env[61906]: ERROR nova.scheduler.client.report [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [req-73d1a830-11f2-4dca-84b8-31d914a9bc18] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID a5b57df7-a16e-44f9-9b2d-23c518860263. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-73d1a830-11f2-4dca-84b8-31d914a9bc18"}]} [ 1016.810708] env[61906]: DEBUG nova.scheduler.client.report [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Refreshing inventories for resource provider a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1016.828566] env[61906]: DEBUG nova.scheduler.client.report [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Updating ProviderTree inventory for provider a5b57df7-a16e-44f9-9b2d-23c518860263 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1016.829049] env[61906]: DEBUG nova.compute.provider_tree [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 130, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1016.840914] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.841986] env[61906]: DEBUG nova.scheduler.client.report [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Refreshing aggregate associations for resource provider a5b57df7-a16e-44f9-9b2d-23c518860263, aggregates: None {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1016.860521] env[61906]: DEBUG nova.scheduler.client.report [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Refreshing trait associations for resource provider a5b57df7-a16e-44f9-9b2d-23c518860263, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64 {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1016.901362] env[61906]: DEBUG oslo_vmware.api [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333587, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.281796} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.901638] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1016.901828] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1016.902014] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1016.902205] env[61906]: INFO nova.compute.manager [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1016.902623] env[61906]: DEBUG oslo.service.loopingcall [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1016.902623] env[61906]: DEBUG nova.compute.manager [-] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1016.902756] env[61906]: DEBUG nova.network.neutron [-] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1016.969780] env[61906]: DEBUG oslo_concurrency.lockutils [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Releasing lock "refresh_cache-a3e3ad00-6921-4072-8cb1-d80302883513" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.969780] env[61906]: DEBUG nova.objects.instance [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lazy-loading 'migration_context' on Instance uuid a3e3ad00-6921-4072-8cb1-d80302883513 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1017.031560] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d1888de-443f-49d9-a786-60e506060f0f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.039535] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e485dd16-1556-4312-9b57-8824d77796da {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.073356] env[61906]: DEBUG nova.network.neutron [req-5fd47e39-a05f-4510-91c2-5d79a3c503c0 req-52a2a730-1702-4248-b049-a6ea690b6b52 service nova] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Updated VIF entry in instance network info cache for port 99bfd128-61db-4547-8a4f-2f3d676352f1. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1017.073919] env[61906]: DEBUG nova.network.neutron [req-5fd47e39-a05f-4510-91c2-5d79a3c503c0 req-52a2a730-1702-4248-b049-a6ea690b6b52 service nova] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Updating instance_info_cache with network_info: [{"id": "99bfd128-61db-4547-8a4f-2f3d676352f1", "address": "fa:16:3e:4a:44:67", "network": {"id": "bc9185e9-f26b-489c-84ee-e6db41ac6283", "bridge": "br-int", "label": "tempest-ServersTestJSON-1167747148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53644a4143a24759a9ff2b5e28b84fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99bfd128-61", "ovs_interfaceid": "99bfd128-61db-4547-8a4f-2f3d676352f1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.076758] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0b6f734-4a0d-4310-ba2a-440a54645c79 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.089091] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaeaf4a3-7e05-4450-b28e-54d5b6f390bd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.104668] env[61906]: DEBUG nova.compute.provider_tree [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1017.284062] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333585, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.607319} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.284062] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] df605d77-7766-47b3-bd27-97361ff8ae4d/df605d77-7766-47b3-bd27-97361ff8ae4d.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1017.284304] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1017.286581] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-22c0ddf8-323c-4a88-a4ca-b84679a470fe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.294170] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 1017.294170] env[61906]: value = "task-1333588" [ 1017.294170] env[61906]: _type = "Task" [ 1017.294170] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.299182] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.304271] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333588, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.473118] env[61906]: DEBUG nova.objects.base [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1017.474474] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b50bdab2-cf27-4962-aaf8-1b22d3339905 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.494966] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f548e25-dec3-49b3-92df-13071c61be7f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.502252] env[61906]: DEBUG oslo_vmware.api [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 1017.502252] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5272ccfa-d9cd-c4fd-7cf9-f28f50c906a3" [ 1017.502252] env[61906]: _type = "Task" [ 1017.502252] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.511048] env[61906]: DEBUG oslo_vmware.api [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5272ccfa-d9cd-c4fd-7cf9-f28f50c906a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.583222] env[61906]: DEBUG oslo_concurrency.lockutils [req-5fd47e39-a05f-4510-91c2-5d79a3c503c0 req-52a2a730-1702-4248-b049-a6ea690b6b52 service nova] Releasing lock "refresh_cache-df605d77-7766-47b3-bd27-97361ff8ae4d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.637455] env[61906]: DEBUG nova.scheduler.client.report [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Updated inventory for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with generation 117 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1017.637747] env[61906]: DEBUG nova.compute.provider_tree [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Updating resource provider a5b57df7-a16e-44f9-9b2d-23c518860263 generation from 117 to 118 during operation: update_inventory {{(pid=61906) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1017.637938] env[61906]: DEBUG nova.compute.provider_tree [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1017.665608] env[61906]: DEBUG nova.network.neutron [-] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.806025] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333588, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.213562} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.806314] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1017.807087] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-146c411a-6260-47b2-a106-8d65dbddb35d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.828495] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] df605d77-7766-47b3-bd27-97361ff8ae4d/df605d77-7766-47b3-bd27-97361ff8ae4d.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1017.828746] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-180873c2-a6b0-488b-ba85-81153805c222 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.849943] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 1017.849943] env[61906]: value = "task-1333589" [ 1017.849943] env[61906]: _type = "Task" [ 1017.849943] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.858851] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333589, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.974962] env[61906]: DEBUG nova.compute.manager [req-278c2187-16c7-485a-8428-b77725eab460 req-9526b86d-ca87-4012-bde1-e4b6f4e376a1 service nova] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Received event network-vif-deleted-b5eb6e7b-4b1c-411f-8d99-ad4888193eda {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1018.012794] env[61906]: DEBUG oslo_vmware.api [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5272ccfa-d9cd-c4fd-7cf9-f28f50c906a3, 'name': SearchDatastore_Task, 'duration_secs': 0.009601} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.013176] env[61906]: DEBUG oslo_concurrency.lockutils [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.168236] env[61906]: INFO nova.compute.manager [-] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Took 1.27 seconds to deallocate network for instance. [ 1018.360831] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333589, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.648563] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 3.161s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.651510] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.811s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.652271] env[61906]: DEBUG nova.objects.instance [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lazy-loading 'resources' on Instance uuid 7504dfa6-bedc-4701-b4fc-60e19e742276 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.674286] env[61906]: DEBUG oslo_concurrency.lockutils [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.869254] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333589, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.211381] env[61906]: INFO nova.scheduler.client.report [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Deleted allocation for migration f4990fe8-e805-4280-8df9-576e6ed39d3b [ 1019.292850] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e209ffb-144c-45a6-a247-70a30d99b606 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.301066] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad600f8f-6fb5-42d3-92dc-233e49bfbebc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.332238] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c3f7494-55c3-41e6-9298-89ebc10280ba {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.340074] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab9b1b80-90e5-4cca-9c01-c81145fce403 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.353300] env[61906]: DEBUG nova.compute.provider_tree [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1019.362892] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333589, 'name': ReconfigVM_Task, 'duration_secs': 1.03445} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.363843] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Reconfigured VM instance instance-0000005f to attach disk [datastore2] df605d77-7766-47b3-bd27-97361ff8ae4d/df605d77-7766-47b3-bd27-97361ff8ae4d.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1019.364528] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-58c911ba-1cff-427e-bb4e-0c2ea755ee91 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.371474] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 1019.371474] env[61906]: value = "task-1333590" [ 1019.371474] env[61906]: _type = "Task" [ 1019.371474] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.381946] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333590, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.718552] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ddd6b217-74fb-4d60-bfb4-c959945b84fa tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "458a1669-a62b-4313-874a-e49809d5c034" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.957s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.856316] env[61906]: DEBUG nova.scheduler.client.report [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1019.882803] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333590, 'name': Rename_Task, 'duration_secs': 0.507917} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.883228] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1019.883559] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0e18e3d5-f008-4599-a049-15ac1fee43e5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.891153] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 1019.891153] env[61906]: value = "task-1333591" [ 1019.891153] env[61906]: _type = "Task" [ 1019.891153] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.902376] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333591, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.185020] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "458a1669-a62b-4313-874a-e49809d5c034" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.185315] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "458a1669-a62b-4313-874a-e49809d5c034" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.185533] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "458a1669-a62b-4313-874a-e49809d5c034-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.185890] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "458a1669-a62b-4313-874a-e49809d5c034-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.186140] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "458a1669-a62b-4313-874a-e49809d5c034-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.188255] env[61906]: INFO nova.compute.manager [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Terminating instance [ 1020.190043] env[61906]: DEBUG nova.compute.manager [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1020.190219] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1020.191072] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4bf33bd-d6ff-4751-b583-f2c09482d6a1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.199519] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1020.199756] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2d8aa2e6-1858-4d31-be5a-65be8ea36b5c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.207038] env[61906]: DEBUG oslo_vmware.api [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 1020.207038] env[61906]: value = "task-1333592" [ 1020.207038] env[61906]: _type = "Task" [ 1020.207038] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.215858] env[61906]: DEBUG oslo_vmware.api [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333592, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.362602] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.711s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.365172] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.066s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.366732] env[61906]: INFO nova.compute.claims [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1020.387774] env[61906]: INFO nova.scheduler.client.report [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Deleted allocations for instance 7504dfa6-bedc-4701-b4fc-60e19e742276 [ 1020.400978] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333591, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.636078] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "ed213a8b-7d7e-47e1-8a4f-d657c13df3ce" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.636321] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "ed213a8b-7d7e-47e1-8a4f-d657c13df3ce" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.717310] env[61906]: DEBUG oslo_vmware.api [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333592, 'name': PowerOffVM_Task, 'duration_secs': 0.174144} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.717578] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1020.717752] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1020.718022] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6b123098-e6d6-41d2-87b8-e3340e982b25 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.779112] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1020.779539] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1020.779614] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Deleting the datastore file [datastore1] 458a1669-a62b-4313-874a-e49809d5c034 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1020.779822] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc5a10f1-9513-463e-b8bf-ee69338f2546 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.786732] env[61906]: DEBUG oslo_vmware.api [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 1020.786732] env[61906]: value = "task-1333594" [ 1020.786732] env[61906]: _type = "Task" [ 1020.786732] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.794610] env[61906]: DEBUG oslo_vmware.api [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333594, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.898293] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4b6795cc-6e22-4992-b3fc-ae5eedb773fb tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "7504dfa6-bedc-4701-b4fc-60e19e742276" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.306s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.905712] env[61906]: DEBUG oslo_vmware.api [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333591, 'name': PowerOnVM_Task, 'duration_secs': 0.517883} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.905974] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1020.906193] env[61906]: INFO nova.compute.manager [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Took 9.94 seconds to spawn the instance on the hypervisor. [ 1020.906443] env[61906]: DEBUG nova.compute.manager [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1020.907187] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bc988ed-4c0d-487b-97bd-cb4247eaa2ea {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.139108] env[61906]: DEBUG nova.compute.manager [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1021.298069] env[61906]: DEBUG oslo_vmware.api [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333594, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145661} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.298372] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1021.298567] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1021.298748] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1021.298975] env[61906]: INFO nova.compute.manager [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1021.299283] env[61906]: DEBUG oslo.service.loopingcall [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1021.299491] env[61906]: DEBUG nova.compute.manager [-] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1021.299591] env[61906]: DEBUG nova.network.neutron [-] [instance: 458a1669-a62b-4313-874a-e49809d5c034] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1021.428653] env[61906]: INFO nova.compute.manager [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Took 16.33 seconds to build instance. [ 1021.540520] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aac87151-4d31-43ce-adea-43c883719140 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.551495] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-380bfc64-fc3a-406e-a288-b3e073cfb988 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.589228] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7d575ba-2194-45a7-a2e8-3c605f6d6ce4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.598707] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa92a5ab-70d5-4692-9acb-fad3586b3720 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.615319] env[61906]: DEBUG nova.compute.provider_tree [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1021.654327] env[61906]: DEBUG nova.compute.manager [req-c5fb1600-b010-442e-a303-41219868ff88 req-3ef17dee-4e0b-4d32-b426-f4bf2eba9525 service nova] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Received event network-vif-deleted-da64f0e8-39b7-4d88-aa6d-4ac0939a0a20 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1021.654561] env[61906]: INFO nova.compute.manager [req-c5fb1600-b010-442e-a303-41219868ff88 req-3ef17dee-4e0b-4d32-b426-f4bf2eba9525 service nova] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Neutron deleted interface da64f0e8-39b7-4d88-aa6d-4ac0939a0a20; detaching it from the instance and deleting it from the info cache [ 1021.654748] env[61906]: DEBUG nova.network.neutron [req-c5fb1600-b010-442e-a303-41219868ff88 req-3ef17dee-4e0b-4d32-b426-f4bf2eba9525 service nova] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.660007] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.931528] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6967d4a0-ab4d-429c-813f-e6ebc5a6625d tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "df605d77-7766-47b3-bd27-97361ff8ae4d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.840s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.119300] env[61906]: DEBUG nova.scheduler.client.report [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1022.130977] env[61906]: DEBUG nova.network.neutron [-] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.157251] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5877c4ad-f604-48ac-9411-4413088b75bb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.172562] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b90565be-2e49-4485-883f-133136982f2d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.217422] env[61906]: DEBUG nova.compute.manager [req-c5fb1600-b010-442e-a303-41219868ff88 req-3ef17dee-4e0b-4d32-b426-f4bf2eba9525 service nova] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Detach interface failed, port_id=da64f0e8-39b7-4d88-aa6d-4ac0939a0a20, reason: Instance 458a1669-a62b-4313-874a-e49809d5c034 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1022.625803] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.261s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.626412] env[61906]: DEBUG nova.compute.manager [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1022.629087] env[61906]: DEBUG oslo_concurrency.lockutils [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 4.616s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.633241] env[61906]: INFO nova.compute.manager [-] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Took 1.33 seconds to deallocate network for instance. [ 1022.655022] env[61906]: DEBUG oslo_concurrency.lockutils [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "df605d77-7766-47b3-bd27-97361ff8ae4d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.655022] env[61906]: DEBUG oslo_concurrency.lockutils [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "df605d77-7766-47b3-bd27-97361ff8ae4d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.655022] env[61906]: DEBUG oslo_concurrency.lockutils [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "df605d77-7766-47b3-bd27-97361ff8ae4d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.655022] env[61906]: DEBUG oslo_concurrency.lockutils [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "df605d77-7766-47b3-bd27-97361ff8ae4d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.655022] env[61906]: DEBUG oslo_concurrency.lockutils [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "df605d77-7766-47b3-bd27-97361ff8ae4d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.660628] env[61906]: INFO nova.compute.manager [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Terminating instance [ 1022.662678] env[61906]: DEBUG nova.compute.manager [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1022.662809] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1022.664041] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e77986f-dcf8-4eb3-b63b-6e52dad395e8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.672469] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1022.672740] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-25630636-2180-46ac-8e41-e45418f602a4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.680493] env[61906]: DEBUG oslo_vmware.api [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 1022.680493] env[61906]: value = "task-1333595" [ 1022.680493] env[61906]: _type = "Task" [ 1022.680493] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.689655] env[61906]: DEBUG oslo_vmware.api [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333595, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.133401] env[61906]: DEBUG nova.compute.utils [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1023.137488] env[61906]: DEBUG nova.compute.manager [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1023.137660] env[61906]: DEBUG nova.network.neutron [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1023.140156] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.185689] env[61906]: DEBUG nova.policy [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc6b374325bc4ef9882e5f84ddd62fd1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f2d2198383d74dac90aa9727a62176df', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1023.195777] env[61906]: DEBUG oslo_vmware.api [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333595, 'name': PowerOffVM_Task, 'duration_secs': 0.200396} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.196074] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1023.196256] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1023.196523] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1e12ad20-60da-4544-8f16-e73c449e694f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.256052] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "1be91e16-1acb-4f13-9b26-cae496a0efb8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.256347] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "1be91e16-1acb-4f13-9b26-cae496a0efb8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.279955] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1023.279955] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1023.279955] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Deleting the datastore file [datastore2] df605d77-7766-47b3-bd27-97361ff8ae4d {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1023.279955] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a5d0e4d0-10ae-4cad-b8c6-adabd2243888 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.287817] env[61906]: DEBUG oslo_vmware.api [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 1023.287817] env[61906]: value = "task-1333597" [ 1023.287817] env[61906]: _type = "Task" [ 1023.287817] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.299925] env[61906]: DEBUG oslo_vmware.api [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333597, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.304861] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08ef44af-18b5-4687-9db7-63b35a0c4889 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.312361] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf5b328c-d77a-4ffb-b136-553182db88d5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.345383] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ba19380-1bd1-44cc-999a-edc216984e07 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.354586] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96d0e24d-82ba-4c2c-9dff-44e2b6c4214d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.368683] env[61906]: DEBUG nova.compute.provider_tree [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1023.456384] env[61906]: DEBUG nova.network.neutron [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Successfully created port: 15a5be02-4db5-4752-ace0-a2e1bcec23c7 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1023.638848] env[61906]: DEBUG nova.compute.manager [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1023.759786] env[61906]: DEBUG nova.compute.manager [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1023.798578] env[61906]: DEBUG oslo_vmware.api [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333597, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180878} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.798847] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1023.799050] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1023.799240] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1023.799417] env[61906]: INFO nova.compute.manager [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1023.799661] env[61906]: DEBUG oslo.service.loopingcall [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1023.799849] env[61906]: DEBUG nova.compute.manager [-] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1023.799946] env[61906]: DEBUG nova.network.neutron [-] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1023.871901] env[61906]: DEBUG nova.scheduler.client.report [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1024.012667] env[61906]: DEBUG nova.compute.manager [req-7c6a51a4-f370-4844-970a-3da0a4bcadc6 req-d758f9b0-66a5-4b50-a0a0-186182d0f14c service nova] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Received event network-vif-deleted-99bfd128-61db-4547-8a4f-2f3d676352f1 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1024.012974] env[61906]: INFO nova.compute.manager [req-7c6a51a4-f370-4844-970a-3da0a4bcadc6 req-d758f9b0-66a5-4b50-a0a0-186182d0f14c service nova] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Neutron deleted interface 99bfd128-61db-4547-8a4f-2f3d676352f1; detaching it from the instance and deleting it from the info cache [ 1024.013178] env[61906]: DEBUG nova.network.neutron [req-7c6a51a4-f370-4844-970a-3da0a4bcadc6 req-d758f9b0-66a5-4b50-a0a0-186182d0f14c service nova] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.281591] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.492074] env[61906]: DEBUG nova.network.neutron [-] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.515991] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-32aa264b-a70d-4215-8f25-4e53797f36b3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.529342] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a07bc845-bc8f-4b8e-99a4-c32807c4ad53 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.560643] env[61906]: DEBUG nova.compute.manager [req-7c6a51a4-f370-4844-970a-3da0a4bcadc6 req-d758f9b0-66a5-4b50-a0a0-186182d0f14c service nova] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Detach interface failed, port_id=99bfd128-61db-4547-8a4f-2f3d676352f1, reason: Instance df605d77-7766-47b3-bd27-97361ff8ae4d could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1024.648965] env[61906]: DEBUG nova.compute.manager [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1024.672635] env[61906]: DEBUG nova.virt.hardware [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1024.672885] env[61906]: DEBUG nova.virt.hardware [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1024.673091] env[61906]: DEBUG nova.virt.hardware [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1024.673305] env[61906]: DEBUG nova.virt.hardware [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1024.673454] env[61906]: DEBUG nova.virt.hardware [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1024.673605] env[61906]: DEBUG nova.virt.hardware [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1024.673815] env[61906]: DEBUG nova.virt.hardware [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1024.673978] env[61906]: DEBUG nova.virt.hardware [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1024.674165] env[61906]: DEBUG nova.virt.hardware [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1024.674330] env[61906]: DEBUG nova.virt.hardware [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1024.674507] env[61906]: DEBUG nova.virt.hardware [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1024.675519] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-017a4a2e-0253-4a9f-a2be-61baa631bb66 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.683639] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-428ca052-208f-4bec-b70a-dbbbcd57b842 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.882089] env[61906]: DEBUG oslo_concurrency.lockutils [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.253s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.885107] env[61906]: DEBUG oslo_concurrency.lockutils [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.211s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.885342] env[61906]: DEBUG nova.objects.instance [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lazy-loading 'resources' on Instance uuid 03d3fb1b-0e46-4544-b01d-498a2baf3b45 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1024.979238] env[61906]: DEBUG nova.network.neutron [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Successfully updated port: 15a5be02-4db5-4752-ace0-a2e1bcec23c7 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1024.994740] env[61906]: INFO nova.compute.manager [-] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Took 1.19 seconds to deallocate network for instance. [ 1025.374713] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "interface-7264af54-cd10-4507-8ae3-5bb8a5cd127f-a701c10b-b79f-4f4e-a4e0-7d702211ffab" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.375165] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "interface-7264af54-cd10-4507-8ae3-5bb8a5cd127f-a701c10b-b79f-4f4e-a4e0-7d702211ffab" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.375361] env[61906]: DEBUG nova.objects.instance [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lazy-loading 'flavor' on Instance uuid 7264af54-cd10-4507-8ae3-5bb8a5cd127f {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1025.444682] env[61906]: INFO nova.scheduler.client.report [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Deleted allocation for migration ae34f086-21e9-4cfa-ba48-060c083baa74 [ 1025.481733] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "refresh_cache-0fe121ad-a539-4c04-bb65-b524cb3d91a8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.481790] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquired lock "refresh_cache-0fe121ad-a539-4c04-bb65-b524cb3d91a8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.482048] env[61906]: DEBUG nova.network.neutron [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1025.501537] env[61906]: DEBUG oslo_concurrency.lockutils [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.525289] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-621dfb1e-b007-4c14-9b3b-9963bcf1dec1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.533447] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1b00c79-421e-46af-bc9a-6d7fb716663a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.565399] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df4653d7-1c6f-44fa-b5de-d455c00fa1a7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.572899] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3f4ec91-42b2-49bd-afc2-05f2c188d2ca {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.586126] env[61906]: DEBUG nova.compute.provider_tree [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1025.951909] env[61906]: DEBUG oslo_concurrency.lockutils [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "a3e3ad00-6921-4072-8cb1-d80302883513" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 11.337s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.966891] env[61906]: DEBUG nova.objects.instance [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lazy-loading 'pci_requests' on Instance uuid 7264af54-cd10-4507-8ae3-5bb8a5cd127f {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.010211] env[61906]: DEBUG nova.network.neutron [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1026.089792] env[61906]: DEBUG nova.scheduler.client.report [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1026.132781] env[61906]: DEBUG nova.network.neutron [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Updating instance_info_cache with network_info: [{"id": "15a5be02-4db5-4752-ace0-a2e1bcec23c7", "address": "fa:16:3e:6d:39:4d", "network": {"id": "d123fdf4-6bb4-400b-a0bb-9d477672af2c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2099881656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d2198383d74dac90aa9727a62176df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15a5be02-4d", "ovs_interfaceid": "15a5be02-4db5-4752-ace0-a2e1bcec23c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.470207] env[61906]: DEBUG nova.objects.base [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Object Instance<7264af54-cd10-4507-8ae3-5bb8a5cd127f> lazy-loaded attributes: flavor,pci_requests {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1026.470586] env[61906]: DEBUG nova.network.neutron [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1026.533104] env[61906]: DEBUG nova.policy [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0de9abfd085f4a43a878da8448615bc4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e36ff7a500ef444cbf5a168c5a48208b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1026.595304] env[61906]: DEBUG oslo_concurrency.lockutils [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.710s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.597735] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.938s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.599377] env[61906]: INFO nova.compute.claims [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1026.614163] env[61906]: INFO nova.scheduler.client.report [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Deleted allocations for instance 03d3fb1b-0e46-4544-b01d-498a2baf3b45 [ 1026.635465] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Releasing lock "refresh_cache-0fe121ad-a539-4c04-bb65-b524cb3d91a8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.635802] env[61906]: DEBUG nova.compute.manager [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Instance network_info: |[{"id": "15a5be02-4db5-4752-ace0-a2e1bcec23c7", "address": "fa:16:3e:6d:39:4d", "network": {"id": "d123fdf4-6bb4-400b-a0bb-9d477672af2c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2099881656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d2198383d74dac90aa9727a62176df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15a5be02-4d", "ovs_interfaceid": "15a5be02-4db5-4752-ace0-a2e1bcec23c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1026.636234] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:39:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cba18f15-a919-422e-a423-1e705e233389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '15a5be02-4db5-4752-ace0-a2e1bcec23c7', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1026.646456] env[61906]: DEBUG oslo.service.loopingcall [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1026.648508] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1026.652109] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-30ca832a-8ef5-4ae8-93a5-118e0eace8fe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.676110] env[61906]: DEBUG nova.compute.manager [req-988c685e-bb44-4e56-bc63-1b0408e99bfe req-70db3e9d-423a-4d4b-a91c-ee861053d226 service nova] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Received event network-vif-plugged-15a5be02-4db5-4752-ace0-a2e1bcec23c7 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1026.676648] env[61906]: DEBUG oslo_concurrency.lockutils [req-988c685e-bb44-4e56-bc63-1b0408e99bfe req-70db3e9d-423a-4d4b-a91c-ee861053d226 service nova] Acquiring lock "0fe121ad-a539-4c04-bb65-b524cb3d91a8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.676935] env[61906]: DEBUG oslo_concurrency.lockutils [req-988c685e-bb44-4e56-bc63-1b0408e99bfe req-70db3e9d-423a-4d4b-a91c-ee861053d226 service nova] Lock "0fe121ad-a539-4c04-bb65-b524cb3d91a8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.677136] env[61906]: DEBUG oslo_concurrency.lockutils [req-988c685e-bb44-4e56-bc63-1b0408e99bfe req-70db3e9d-423a-4d4b-a91c-ee861053d226 service nova] Lock "0fe121ad-a539-4c04-bb65-b524cb3d91a8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.677324] env[61906]: DEBUG nova.compute.manager [req-988c685e-bb44-4e56-bc63-1b0408e99bfe req-70db3e9d-423a-4d4b-a91c-ee861053d226 service nova] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] No waiting events found dispatching network-vif-plugged-15a5be02-4db5-4752-ace0-a2e1bcec23c7 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1026.677511] env[61906]: WARNING nova.compute.manager [req-988c685e-bb44-4e56-bc63-1b0408e99bfe req-70db3e9d-423a-4d4b-a91c-ee861053d226 service nova] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Received unexpected event network-vif-plugged-15a5be02-4db5-4752-ace0-a2e1bcec23c7 for instance with vm_state building and task_state spawning. [ 1026.679494] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1026.679494] env[61906]: value = "task-1333598" [ 1026.679494] env[61906]: _type = "Task" [ 1026.679494] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.690568] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333598, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.122305] env[61906]: DEBUG oslo_concurrency.lockutils [None req-63a601d2-9d76-4919-b968-585f576196ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "03d3fb1b-0e46-4544-b01d-498a2baf3b45" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.352s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.151830] env[61906]: DEBUG oslo_concurrency.lockutils [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "a3e3ad00-6921-4072-8cb1-d80302883513" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.152025] env[61906]: DEBUG oslo_concurrency.lockutils [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "a3e3ad00-6921-4072-8cb1-d80302883513" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.152201] env[61906]: DEBUG oslo_concurrency.lockutils [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "a3e3ad00-6921-4072-8cb1-d80302883513-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.152410] env[61906]: DEBUG oslo_concurrency.lockutils [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "a3e3ad00-6921-4072-8cb1-d80302883513-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.152585] env[61906]: DEBUG oslo_concurrency.lockutils [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "a3e3ad00-6921-4072-8cb1-d80302883513-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.154909] env[61906]: INFO nova.compute.manager [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Terminating instance [ 1027.156578] env[61906]: DEBUG nova.compute.manager [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1027.156776] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1027.157638] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b828830-b379-4d88-8bf8-50a0c77cc05f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.165909] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1027.166159] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-09602ac0-6599-48d3-87b5-de3a044bb633 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.173430] env[61906]: DEBUG oslo_vmware.api [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 1027.173430] env[61906]: value = "task-1333599" [ 1027.173430] env[61906]: _type = "Task" [ 1027.173430] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.181673] env[61906]: DEBUG oslo_vmware.api [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333599, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.193315] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333598, 'name': CreateVM_Task, 'duration_secs': 0.314009} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.193516] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1027.194161] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.194357] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.194686] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1027.194938] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a84e0bcb-dd61-44b0-9be6-c591381133ea {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.199612] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 1027.199612] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]524fa888-85ac-02c7-f18e-6a575313d297" [ 1027.199612] env[61906]: _type = "Task" [ 1027.199612] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.207765] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]524fa888-85ac-02c7-f18e-6a575313d297, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.683165] env[61906]: DEBUG oslo_vmware.api [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333599, 'name': PowerOffVM_Task, 'duration_secs': 0.180035} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.685589] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1027.685780] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1027.686217] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d2c19aff-2152-422f-99d7-d67bb88f78c3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.710501] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]524fa888-85ac-02c7-f18e-6a575313d297, 'name': SearchDatastore_Task, 'duration_secs': 0.010166} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.710679] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.711294] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1027.711294] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.711294] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.711478] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1027.711605] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7e683a4a-a1f8-48a6-a512-afcd09622208 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.720040] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1027.720244] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1027.723213] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ebd7cb31-7c77-4f66-9614-95fe1c669c31 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.728522] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 1027.728522] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]526100ef-f55e-c753-8ce8-4e0f23c773a1" [ 1027.728522] env[61906]: _type = "Task" [ 1027.728522] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.736239] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]526100ef-f55e-c753-8ce8-4e0f23c773a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.747474] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a272b5aa-b6af-4505-a721-aaa1a3e64ff8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.753983] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-377ee7a8-2442-4b46-935c-98fe776b0d2b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.785592] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60770e40-571e-4d03-b166-863f87ac8452 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.793667] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63d2f2b9-661d-4861-a5a7-b8eac623b63b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.807366] env[61906]: DEBUG nova.compute.provider_tree [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1027.998709] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1027.998921] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1027.999025] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Deleting the datastore file [datastore1] a3e3ad00-6921-4072-8cb1-d80302883513 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1027.999324] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e949b1c4-b817-43a7-82ab-0591b17db0d8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.006875] env[61906]: DEBUG oslo_vmware.api [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for the task: (returnval){ [ 1028.006875] env[61906]: value = "task-1333601" [ 1028.006875] env[61906]: _type = "Task" [ 1028.006875] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.015054] env[61906]: DEBUG oslo_vmware.api [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333601, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.239439] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]526100ef-f55e-c753-8ce8-4e0f23c773a1, 'name': SearchDatastore_Task, 'duration_secs': 0.00914} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.240635] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40fd9689-471b-4196-8599-1ec6e365910b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.246767] env[61906]: DEBUG nova.network.neutron [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Successfully updated port: a701c10b-b79f-4f4e-a4e0-7d702211ffab {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1028.251800] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 1028.251800] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]526f3aaf-72d3-274a-8588-f2f9a0383baf" [ 1028.251800] env[61906]: _type = "Task" [ 1028.251800] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.263748] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]526f3aaf-72d3-274a-8588-f2f9a0383baf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.310693] env[61906]: DEBUG nova.scheduler.client.report [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1028.516589] env[61906]: DEBUG oslo_vmware.api [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Task: {'id': task-1333601, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145098} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.516825] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1028.516995] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1028.517194] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1028.517371] env[61906]: INFO nova.compute.manager [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Took 1.36 seconds to destroy the instance on the hypervisor. [ 1028.517615] env[61906]: DEBUG oslo.service.loopingcall [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1028.517807] env[61906]: DEBUG nova.compute.manager [-] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1028.517903] env[61906]: DEBUG nova.network.neutron [-] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1028.711199] env[61906]: DEBUG nova.compute.manager [req-a83e910a-cdb2-4d91-aa5c-c9901af33339 req-2cd9d07d-9d91-41fc-ac2d-17d1ba2aedfa service nova] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Received event network-changed-15a5be02-4db5-4752-ace0-a2e1bcec23c7 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1028.711199] env[61906]: DEBUG nova.compute.manager [req-a83e910a-cdb2-4d91-aa5c-c9901af33339 req-2cd9d07d-9d91-41fc-ac2d-17d1ba2aedfa service nova] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Refreshing instance network info cache due to event network-changed-15a5be02-4db5-4752-ace0-a2e1bcec23c7. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1028.711199] env[61906]: DEBUG oslo_concurrency.lockutils [req-a83e910a-cdb2-4d91-aa5c-c9901af33339 req-2cd9d07d-9d91-41fc-ac2d-17d1ba2aedfa service nova] Acquiring lock "refresh_cache-0fe121ad-a539-4c04-bb65-b524cb3d91a8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.711505] env[61906]: DEBUG oslo_concurrency.lockutils [req-a83e910a-cdb2-4d91-aa5c-c9901af33339 req-2cd9d07d-9d91-41fc-ac2d-17d1ba2aedfa service nova] Acquired lock "refresh_cache-0fe121ad-a539-4c04-bb65-b524cb3d91a8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.711505] env[61906]: DEBUG nova.network.neutron [req-a83e910a-cdb2-4d91-aa5c-c9901af33339 req-2cd9d07d-9d91-41fc-ac2d-17d1ba2aedfa service nova] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Refreshing network info cache for port 15a5be02-4db5-4752-ace0-a2e1bcec23c7 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1028.749864] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.750075] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.750263] env[61906]: DEBUG nova.network.neutron [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1028.762983] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]526f3aaf-72d3-274a-8588-f2f9a0383baf, 'name': SearchDatastore_Task, 'duration_secs': 0.013013} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.763272] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.763525] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 0fe121ad-a539-4c04-bb65-b524cb3d91a8/0fe121ad-a539-4c04-bb65-b524cb3d91a8.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1028.763779] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7c37759e-c0f0-4585-9373-0f0f1621f5f8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.771268] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 1028.771268] env[61906]: value = "task-1333602" [ 1028.771268] env[61906]: _type = "Task" [ 1028.771268] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.779862] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333602, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.815363] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.218s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.815958] env[61906]: DEBUG nova.compute.manager [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1028.819090] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.679s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.819311] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.821712] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.540s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.823479] env[61906]: INFO nova.compute.claims [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1028.848868] env[61906]: INFO nova.scheduler.client.report [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Deleted allocations for instance 458a1669-a62b-4313-874a-e49809d5c034 [ 1029.239128] env[61906]: DEBUG nova.network.neutron [-] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.283483] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333602, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.287679] env[61906]: WARNING nova.network.neutron [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] 37d8737f-db44-4cf6-8e34-31b4cbd2f82e already exists in list: networks containing: ['37d8737f-db44-4cf6-8e34-31b4cbd2f82e']. ignoring it [ 1029.320453] env[61906]: DEBUG nova.compute.utils [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1029.323951] env[61906]: DEBUG nova.compute.manager [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1029.324279] env[61906]: DEBUG nova.network.neutron [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1029.357128] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cc929cc1-e9ce-45db-8095-3b23eebe1c38 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "458a1669-a62b-4313-874a-e49809d5c034" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.172s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.408021] env[61906]: DEBUG nova.policy [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bfb8ebd8f94f45dfadb1d8802d04aec9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4f7fc2042e6549cabe98838990f35842', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1029.431132] env[61906]: DEBUG nova.network.neutron [req-a83e910a-cdb2-4d91-aa5c-c9901af33339 req-2cd9d07d-9d91-41fc-ac2d-17d1ba2aedfa service nova] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Updated VIF entry in instance network info cache for port 15a5be02-4db5-4752-ace0-a2e1bcec23c7. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1029.431525] env[61906]: DEBUG nova.network.neutron [req-a83e910a-cdb2-4d91-aa5c-c9901af33339 req-2cd9d07d-9d91-41fc-ac2d-17d1ba2aedfa service nova] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Updating instance_info_cache with network_info: [{"id": "15a5be02-4db5-4752-ace0-a2e1bcec23c7", "address": "fa:16:3e:6d:39:4d", "network": {"id": "d123fdf4-6bb4-400b-a0bb-9d477672af2c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2099881656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f2d2198383d74dac90aa9727a62176df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15a5be02-4d", "ovs_interfaceid": "15a5be02-4db5-4752-ace0-a2e1bcec23c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.742131] env[61906]: INFO nova.compute.manager [-] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Took 1.22 seconds to deallocate network for instance. [ 1029.782310] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333602, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.555833} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.782617] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 0fe121ad-a539-4c04-bb65-b524cb3d91a8/0fe121ad-a539-4c04-bb65-b524cb3d91a8.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1029.782861] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1029.783156] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1e7c674d-3839-42a5-a581-d76335af0c7a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.791918] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 1029.791918] env[61906]: value = "task-1333603" [ 1029.791918] env[61906]: _type = "Task" [ 1029.791918] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.792949] env[61906]: DEBUG nova.network.neutron [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Updating instance_info_cache with network_info: [{"id": "73d0e617-fd5e-4fa2-81c5-710a3beac1a4", "address": "fa:16:3e:66:bf:d9", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73d0e617-fd", "ovs_interfaceid": "73d0e617-fd5e-4fa2-81c5-710a3beac1a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a701c10b-b79f-4f4e-a4e0-7d702211ffab", "address": "fa:16:3e:d7:47:ab", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa701c10b-b7", "ovs_interfaceid": "a701c10b-b79f-4f4e-a4e0-7d702211ffab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.805017] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333603, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.828078] env[61906]: DEBUG nova.compute.manager [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1029.837049] env[61906]: DEBUG nova.network.neutron [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Successfully created port: e55c6ef7-fa51-455b-a1da-2fea0facd0a2 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1029.934524] env[61906]: DEBUG oslo_concurrency.lockutils [req-a83e910a-cdb2-4d91-aa5c-c9901af33339 req-2cd9d07d-9d91-41fc-ac2d-17d1ba2aedfa service nova] Releasing lock "refresh_cache-0fe121ad-a539-4c04-bb65-b524cb3d91a8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.934524] env[61906]: DEBUG nova.compute.manager [req-a83e910a-cdb2-4d91-aa5c-c9901af33339 req-2cd9d07d-9d91-41fc-ac2d-17d1ba2aedfa service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Received event network-vif-plugged-a701c10b-b79f-4f4e-a4e0-7d702211ffab {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1029.934524] env[61906]: DEBUG oslo_concurrency.lockutils [req-a83e910a-cdb2-4d91-aa5c-c9901af33339 req-2cd9d07d-9d91-41fc-ac2d-17d1ba2aedfa service nova] Acquiring lock "7264af54-cd10-4507-8ae3-5bb8a5cd127f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.934730] env[61906]: DEBUG oslo_concurrency.lockutils [req-a83e910a-cdb2-4d91-aa5c-c9901af33339 req-2cd9d07d-9d91-41fc-ac2d-17d1ba2aedfa service nova] Lock "7264af54-cd10-4507-8ae3-5bb8a5cd127f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.934770] env[61906]: DEBUG oslo_concurrency.lockutils [req-a83e910a-cdb2-4d91-aa5c-c9901af33339 req-2cd9d07d-9d91-41fc-ac2d-17d1ba2aedfa service nova] Lock "7264af54-cd10-4507-8ae3-5bb8a5cd127f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.935567] env[61906]: DEBUG nova.compute.manager [req-a83e910a-cdb2-4d91-aa5c-c9901af33339 req-2cd9d07d-9d91-41fc-ac2d-17d1ba2aedfa service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] No waiting events found dispatching network-vif-plugged-a701c10b-b79f-4f4e-a4e0-7d702211ffab {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1029.935567] env[61906]: WARNING nova.compute.manager [req-a83e910a-cdb2-4d91-aa5c-c9901af33339 req-2cd9d07d-9d91-41fc-ac2d-17d1ba2aedfa service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Received unexpected event network-vif-plugged-a701c10b-b79f-4f4e-a4e0-7d702211ffab for instance with vm_state active and task_state None. [ 1029.935567] env[61906]: DEBUG nova.compute.manager [req-a83e910a-cdb2-4d91-aa5c-c9901af33339 req-2cd9d07d-9d91-41fc-ac2d-17d1ba2aedfa service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Received event network-changed-a701c10b-b79f-4f4e-a4e0-7d702211ffab {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1029.935567] env[61906]: DEBUG nova.compute.manager [req-a83e910a-cdb2-4d91-aa5c-c9901af33339 req-2cd9d07d-9d91-41fc-ac2d-17d1ba2aedfa service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Refreshing instance network info cache due to event network-changed-a701c10b-b79f-4f4e-a4e0-7d702211ffab. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1029.935567] env[61906]: DEBUG oslo_concurrency.lockutils [req-a83e910a-cdb2-4d91-aa5c-c9901af33339 req-2cd9d07d-9d91-41fc-ac2d-17d1ba2aedfa service nova] Acquiring lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.979391] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63e22a97-34b8-468b-89fe-4b82e459d682 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.987957] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92dac5b9-da7f-4eac-9f23-914e4bc7eb2b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.020082] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c2bdd57-f5a4-4e19-8a3d-6eccc3f46b48 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.030370] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0246dad-239c-4c29-a644-2d0b591b0bd5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.045139] env[61906]: DEBUG nova.compute.provider_tree [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1030.249498] env[61906]: DEBUG oslo_concurrency.lockutils [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.296882] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Releasing lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.297547] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.297712] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired lock "7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.297983] env[61906]: DEBUG oslo_concurrency.lockutils [req-a83e910a-cdb2-4d91-aa5c-c9901af33339 req-2cd9d07d-9d91-41fc-ac2d-17d1ba2aedfa service nova] Acquired lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.298182] env[61906]: DEBUG nova.network.neutron [req-a83e910a-cdb2-4d91-aa5c-c9901af33339 req-2cd9d07d-9d91-41fc-ac2d-17d1ba2aedfa service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Refreshing network info cache for port a701c10b-b79f-4f4e-a4e0-7d702211ffab {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1030.299696] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f88eb8d3-955b-46a1-a388-223f2aa8b5ac {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.308386] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333603, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.096956} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.320603] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1030.321673] env[61906]: DEBUG nova.virt.hardware [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1030.321921] env[61906]: DEBUG nova.virt.hardware [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1030.322099] env[61906]: DEBUG nova.virt.hardware [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1030.322292] env[61906]: DEBUG nova.virt.hardware [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1030.322444] env[61906]: DEBUG nova.virt.hardware [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1030.322594] env[61906]: DEBUG nova.virt.hardware [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1030.322800] env[61906]: DEBUG nova.virt.hardware [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1030.322962] env[61906]: DEBUG nova.virt.hardware [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1030.323171] env[61906]: DEBUG nova.virt.hardware [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1030.323345] env[61906]: DEBUG nova.virt.hardware [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1030.323522] env[61906]: DEBUG nova.virt.hardware [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1030.329977] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Reconfiguring VM to attach interface {{(pid=61906) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1030.330757] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6419809d-dfa4-4688-a4b4-902106b312f4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.333577] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c46d161-ecd7-4cae-8189-4da9b4b1641c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.371107] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] 0fe121ad-a539-4c04-bb65-b524cb3d91a8/0fe121ad-a539-4c04-bb65-b524cb3d91a8.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1030.372862] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f81bc04e-f302-46ad-bb3f-cec35502534f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.387255] env[61906]: DEBUG oslo_vmware.api [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 1030.387255] env[61906]: value = "task-1333604" [ 1030.387255] env[61906]: _type = "Task" [ 1030.387255] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.395123] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 1030.395123] env[61906]: value = "task-1333605" [ 1030.395123] env[61906]: _type = "Task" [ 1030.395123] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.399143] env[61906]: DEBUG oslo_vmware.api [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333604, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.406966] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333605, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.548418] env[61906]: DEBUG nova.scheduler.client.report [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1030.737380] env[61906]: DEBUG nova.compute.manager [req-92741cf8-bf1e-4f92-8f6b-d12dd3ed3fb3 req-46913520-b6d8-4762-b989-744c4a60e716 service nova] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Received event network-vif-deleted-e3ce81bf-b27e-47ea-9c48-a17e608a9cd6 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1030.854396] env[61906]: DEBUG nova.compute.manager [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1030.881471] env[61906]: DEBUG nova.virt.hardware [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1030.882187] env[61906]: DEBUG nova.virt.hardware [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1030.882187] env[61906]: DEBUG nova.virt.hardware [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1030.882187] env[61906]: DEBUG nova.virt.hardware [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1030.882426] env[61906]: DEBUG nova.virt.hardware [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1030.882733] env[61906]: DEBUG nova.virt.hardware [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1030.882733] env[61906]: DEBUG nova.virt.hardware [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1030.882876] env[61906]: DEBUG nova.virt.hardware [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1030.883108] env[61906]: DEBUG nova.virt.hardware [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1030.883820] env[61906]: DEBUG nova.virt.hardware [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1030.883820] env[61906]: DEBUG nova.virt.hardware [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1030.884672] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c95cf12-492e-4d6f-b785-f04ceb8e9947 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.897092] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99121a38-d09f-46bb-b7fc-208791f2eb13 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.904574] env[61906]: DEBUG oslo_vmware.api [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333604, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.920180] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333605, 'name': ReconfigVM_Task, 'duration_secs': 0.351232} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.920463] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Reconfigured VM instance instance-00000060 to attach disk [datastore2] 0fe121ad-a539-4c04-bb65-b524cb3d91a8/0fe121ad-a539-4c04-bb65-b524cb3d91a8.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1030.921083] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3f373880-dc9b-4612-9e52-46f71d1e917d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.927601] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 1030.927601] env[61906]: value = "task-1333606" [ 1030.927601] env[61906]: _type = "Task" [ 1030.927601] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.937723] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333606, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.034755] env[61906]: DEBUG nova.network.neutron [req-a83e910a-cdb2-4d91-aa5c-c9901af33339 req-2cd9d07d-9d91-41fc-ac2d-17d1ba2aedfa service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Updated VIF entry in instance network info cache for port a701c10b-b79f-4f4e-a4e0-7d702211ffab. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1031.035249] env[61906]: DEBUG nova.network.neutron [req-a83e910a-cdb2-4d91-aa5c-c9901af33339 req-2cd9d07d-9d91-41fc-ac2d-17d1ba2aedfa service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Updating instance_info_cache with network_info: [{"id": "73d0e617-fd5e-4fa2-81c5-710a3beac1a4", "address": "fa:16:3e:66:bf:d9", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73d0e617-fd", "ovs_interfaceid": "73d0e617-fd5e-4fa2-81c5-710a3beac1a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a701c10b-b79f-4f4e-a4e0-7d702211ffab", "address": "fa:16:3e:d7:47:ab", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa701c10b-b7", "ovs_interfaceid": "a701c10b-b79f-4f4e-a4e0-7d702211ffab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.055134] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.233s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.055769] env[61906]: DEBUG nova.compute.manager [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1031.058708] env[61906]: DEBUG oslo_concurrency.lockutils [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.557s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.058964] env[61906]: DEBUG nova.objects.instance [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lazy-loading 'resources' on Instance uuid df605d77-7766-47b3-bd27-97361ff8ae4d {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.231356] env[61906]: DEBUG nova.compute.manager [req-7841dac5-980c-4f11-99f9-65b10b32e6c9 req-548a432b-6c31-4820-bed3-7cd726caf97d service nova] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Received event network-vif-plugged-e55c6ef7-fa51-455b-a1da-2fea0facd0a2 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1031.231589] env[61906]: DEBUG oslo_concurrency.lockutils [req-7841dac5-980c-4f11-99f9-65b10b32e6c9 req-548a432b-6c31-4820-bed3-7cd726caf97d service nova] Acquiring lock "ed213a8b-7d7e-47e1-8a4f-d657c13df3ce-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.231809] env[61906]: DEBUG oslo_concurrency.lockutils [req-7841dac5-980c-4f11-99f9-65b10b32e6c9 req-548a432b-6c31-4820-bed3-7cd726caf97d service nova] Lock "ed213a8b-7d7e-47e1-8a4f-d657c13df3ce-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.231982] env[61906]: DEBUG oslo_concurrency.lockutils [req-7841dac5-980c-4f11-99f9-65b10b32e6c9 req-548a432b-6c31-4820-bed3-7cd726caf97d service nova] Lock "ed213a8b-7d7e-47e1-8a4f-d657c13df3ce-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.232400] env[61906]: DEBUG nova.compute.manager [req-7841dac5-980c-4f11-99f9-65b10b32e6c9 req-548a432b-6c31-4820-bed3-7cd726caf97d service nova] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] No waiting events found dispatching network-vif-plugged-e55c6ef7-fa51-455b-a1da-2fea0facd0a2 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1031.232664] env[61906]: WARNING nova.compute.manager [req-7841dac5-980c-4f11-99f9-65b10b32e6c9 req-548a432b-6c31-4820-bed3-7cd726caf97d service nova] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Received unexpected event network-vif-plugged-e55c6ef7-fa51-455b-a1da-2fea0facd0a2 for instance with vm_state building and task_state spawning. [ 1031.325855] env[61906]: DEBUG nova.network.neutron [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Successfully updated port: e55c6ef7-fa51-455b-a1da-2fea0facd0a2 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1031.399575] env[61906]: DEBUG oslo_vmware.api [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333604, 'name': ReconfigVM_Task, 'duration_secs': 0.624783} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.400114] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Releasing lock "7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.400345] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Reconfigured VM to attach interface {{(pid=61906) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1031.440241] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333606, 'name': Rename_Task, 'duration_secs': 0.177796} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.440538] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1031.440782] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-12b7da21-cf64-44f3-9006-e878d821533b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.447203] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 1031.447203] env[61906]: value = "task-1333607" [ 1031.447203] env[61906]: _type = "Task" [ 1031.447203] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.455385] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333607, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.537951] env[61906]: DEBUG oslo_concurrency.lockutils [req-a83e910a-cdb2-4d91-aa5c-c9901af33339 req-2cd9d07d-9d91-41fc-ac2d-17d1ba2aedfa service nova] Releasing lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.562438] env[61906]: DEBUG nova.compute.utils [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1031.567060] env[61906]: DEBUG nova.compute.manager [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1031.567215] env[61906]: DEBUG nova.network.neutron [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1031.605373] env[61906]: DEBUG nova.policy [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1b017cdae2af4ae8aa069ae23e744797', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '129cdb06b47346b6895df45e73f7b2dd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1031.687915] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6c22955-70f5-4398-8092-c468cce3785a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.696117] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ccc6e27-3631-4430-83f3-b2881c8cde1a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.727023] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-269b84a7-5499-4a63-bcf0-95c9e75709f1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.735221] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d937333-6d6b-499b-b23b-beae123b4cba {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.749250] env[61906]: DEBUG nova.compute.provider_tree [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1031.828372] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "refresh_cache-ed213a8b-7d7e-47e1-8a4f-d657c13df3ce" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.828455] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired lock "refresh_cache-ed213a8b-7d7e-47e1-8a4f-d657c13df3ce" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.828613] env[61906]: DEBUG nova.network.neutron [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1031.889847] env[61906]: DEBUG nova.network.neutron [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Successfully created port: df3f0533-5c71-429f-9e5b-a39a6eb52ff8 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1031.905100] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2dc7363e-8010-46b8-9783-4c611b94255a tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "interface-7264af54-cd10-4507-8ae3-5bb8a5cd127f-a701c10b-b79f-4f4e-a4e0-7d702211ffab" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.530s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.959285] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333607, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.068581] env[61906]: DEBUG nova.compute.manager [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1032.253360] env[61906]: DEBUG nova.scheduler.client.report [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1032.359609] env[61906]: DEBUG nova.network.neutron [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1032.464435] env[61906]: DEBUG oslo_vmware.api [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333607, 'name': PowerOnVM_Task, 'duration_secs': 0.574854} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.464714] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1032.464920] env[61906]: INFO nova.compute.manager [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Took 7.82 seconds to spawn the instance on the hypervisor. [ 1032.465117] env[61906]: DEBUG nova.compute.manager [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1032.466371] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-303e59c2-9b22-4b40-a001-0d89b9b90676 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.493512] env[61906]: DEBUG nova.network.neutron [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Updating instance_info_cache with network_info: [{"id": "e55c6ef7-fa51-455b-a1da-2fea0facd0a2", "address": "fa:16:3e:86:a7:d3", "network": {"id": "af039bdd-ffd2-483d-9655-e7df48936290", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-111531540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f7fc2042e6549cabe98838990f35842", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape55c6ef7-fa", "ovs_interfaceid": "e55c6ef7-fa51-455b-a1da-2fea0facd0a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.758625] env[61906]: DEBUG oslo_concurrency.lockutils [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.700s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.761653] env[61906]: DEBUG oslo_concurrency.lockutils [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.512s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.762071] env[61906]: DEBUG oslo_concurrency.lockutils [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.785514] env[61906]: INFO nova.scheduler.client.report [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Deleted allocations for instance df605d77-7766-47b3-bd27-97361ff8ae4d [ 1032.787151] env[61906]: INFO nova.scheduler.client.report [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Deleted allocations for instance a3e3ad00-6921-4072-8cb1-d80302883513 [ 1032.985139] env[61906]: INFO nova.compute.manager [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Took 15.70 seconds to build instance. [ 1032.996830] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Releasing lock "refresh_cache-ed213a8b-7d7e-47e1-8a4f-d657c13df3ce" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.997174] env[61906]: DEBUG nova.compute.manager [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Instance network_info: |[{"id": "e55c6ef7-fa51-455b-a1da-2fea0facd0a2", "address": "fa:16:3e:86:a7:d3", "network": {"id": "af039bdd-ffd2-483d-9655-e7df48936290", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-111531540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f7fc2042e6549cabe98838990f35842", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape55c6ef7-fa", "ovs_interfaceid": "e55c6ef7-fa51-455b-a1da-2fea0facd0a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1032.997814] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:86:a7:d3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '19440099-773e-4a31-b82e-84a4daa5d8fe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e55c6ef7-fa51-455b-a1da-2fea0facd0a2', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1033.005508] env[61906]: DEBUG oslo.service.loopingcall [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1033.005959] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1033.006203] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8789e37b-1e81-4348-8e15-1dade5d8bad1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.027530] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1033.027530] env[61906]: value = "task-1333608" [ 1033.027530] env[61906]: _type = "Task" [ 1033.027530] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.077536] env[61906]: DEBUG nova.compute.manager [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1033.104521] env[61906]: DEBUG nova.virt.hardware [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1033.104807] env[61906]: DEBUG nova.virt.hardware [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1033.105066] env[61906]: DEBUG nova.virt.hardware [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1033.105284] env[61906]: DEBUG nova.virt.hardware [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1033.105445] env[61906]: DEBUG nova.virt.hardware [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1033.105598] env[61906]: DEBUG nova.virt.hardware [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1033.105815] env[61906]: DEBUG nova.virt.hardware [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1033.105981] env[61906]: DEBUG nova.virt.hardware [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1033.106176] env[61906]: DEBUG nova.virt.hardware [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1033.106337] env[61906]: DEBUG nova.virt.hardware [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1033.106564] env[61906]: DEBUG nova.virt.hardware [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1033.107666] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c794c6b7-fe29-49e4-9393-cf442062a758 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.119920] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-680777d8-c67a-4f35-9833-d1503da067e0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.260622] env[61906]: DEBUG nova.compute.manager [req-7f160134-a2ce-4a99-a25e-f0f906be0bc5 req-874a5de8-e6bd-453a-8744-990517a431e4 service nova] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Received event network-changed-e55c6ef7-fa51-455b-a1da-2fea0facd0a2 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1033.260780] env[61906]: DEBUG nova.compute.manager [req-7f160134-a2ce-4a99-a25e-f0f906be0bc5 req-874a5de8-e6bd-453a-8744-990517a431e4 service nova] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Refreshing instance network info cache due to event network-changed-e55c6ef7-fa51-455b-a1da-2fea0facd0a2. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1033.260968] env[61906]: DEBUG oslo_concurrency.lockutils [req-7f160134-a2ce-4a99-a25e-f0f906be0bc5 req-874a5de8-e6bd-453a-8744-990517a431e4 service nova] Acquiring lock "refresh_cache-ed213a8b-7d7e-47e1-8a4f-d657c13df3ce" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.261538] env[61906]: DEBUG oslo_concurrency.lockutils [req-7f160134-a2ce-4a99-a25e-f0f906be0bc5 req-874a5de8-e6bd-453a-8744-990517a431e4 service nova] Acquired lock "refresh_cache-ed213a8b-7d7e-47e1-8a4f-d657c13df3ce" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.261834] env[61906]: DEBUG nova.network.neutron [req-7f160134-a2ce-4a99-a25e-f0f906be0bc5 req-874a5de8-e6bd-453a-8744-990517a431e4 service nova] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Refreshing network info cache for port e55c6ef7-fa51-455b-a1da-2fea0facd0a2 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1033.299705] env[61906]: DEBUG oslo_concurrency.lockutils [None req-55ae9892-02b5-4215-addd-4b698a5971a9 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "df605d77-7766-47b3-bd27-97361ff8ae4d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.647s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.304022] env[61906]: DEBUG oslo_concurrency.lockutils [None req-416fe955-43f6-414d-961c-4bfd51826d67 tempest-DeleteServersTestJSON-13404641 tempest-DeleteServersTestJSON-13404641-project-member] Lock "a3e3ad00-6921-4072-8cb1-d80302883513" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.149s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.432976] env[61906]: DEBUG nova.compute.manager [req-fa49a5b8-9dbb-4f26-b8df-e4a3fc617852 req-a8bf5e6b-a5d4-4bab-9130-c1953c8e8cdc service nova] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Received event network-vif-plugged-df3f0533-5c71-429f-9e5b-a39a6eb52ff8 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1033.433302] env[61906]: DEBUG oslo_concurrency.lockutils [req-fa49a5b8-9dbb-4f26-b8df-e4a3fc617852 req-a8bf5e6b-a5d4-4bab-9130-c1953c8e8cdc service nova] Acquiring lock "1be91e16-1acb-4f13-9b26-cae496a0efb8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.433540] env[61906]: DEBUG oslo_concurrency.lockutils [req-fa49a5b8-9dbb-4f26-b8df-e4a3fc617852 req-a8bf5e6b-a5d4-4bab-9130-c1953c8e8cdc service nova] Lock "1be91e16-1acb-4f13-9b26-cae496a0efb8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.433697] env[61906]: DEBUG oslo_concurrency.lockutils [req-fa49a5b8-9dbb-4f26-b8df-e4a3fc617852 req-a8bf5e6b-a5d4-4bab-9130-c1953c8e8cdc service nova] Lock "1be91e16-1acb-4f13-9b26-cae496a0efb8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.433929] env[61906]: DEBUG nova.compute.manager [req-fa49a5b8-9dbb-4f26-b8df-e4a3fc617852 req-a8bf5e6b-a5d4-4bab-9130-c1953c8e8cdc service nova] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] No waiting events found dispatching network-vif-plugged-df3f0533-5c71-429f-9e5b-a39a6eb52ff8 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1033.434145] env[61906]: WARNING nova.compute.manager [req-fa49a5b8-9dbb-4f26-b8df-e4a3fc617852 req-a8bf5e6b-a5d4-4bab-9130-c1953c8e8cdc service nova] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Received unexpected event network-vif-plugged-df3f0533-5c71-429f-9e5b-a39a6eb52ff8 for instance with vm_state building and task_state spawning. [ 1033.487250] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e32653a1-e610-4770-8f94-24ff3feb36ca tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "0fe121ad-a539-4c04-bb65-b524cb3d91a8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.220s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.539637] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333608, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.615629] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "interface-7264af54-cd10-4507-8ae3-5bb8a5cd127f-a701c10b-b79f-4f4e-a4e0-7d702211ffab" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.615942] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "interface-7264af54-cd10-4507-8ae3-5bb8a5cd127f-a701c10b-b79f-4f4e-a4e0-7d702211ffab" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.018089] env[61906]: DEBUG nova.network.neutron [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Successfully updated port: df3f0533-5c71-429f-9e5b-a39a6eb52ff8 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1034.044117] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333608, 'name': CreateVM_Task, 'duration_secs': 0.530921} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.044362] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1034.045128] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.045300] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.045647] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1034.045929] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d55afce-19f3-4035-b823-7d0233592209 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.051327] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 1034.051327] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]527ea359-99ea-df86-fed7-7db07be80760" [ 1034.051327] env[61906]: _type = "Task" [ 1034.051327] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.060261] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]527ea359-99ea-df86-fed7-7db07be80760, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.119110] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.119311] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired lock "7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.120209] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ea4c740-c737-42b6-8a0e-6f0566fda771 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.141151] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f8df2ff-b033-4acb-9a40-b98b505e5a09 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.146290] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "36c6b1a3-be22-4cec-b21c-0445f5988be1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.146512] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "36c6b1a3-be22-4cec-b21c-0445f5988be1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.172653] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Reconfiguring VM to detach interface {{(pid=61906) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1034.173268] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bbd67703-2105-4a03-8e36-8685dfb13929 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.193015] env[61906]: DEBUG oslo_vmware.api [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 1034.193015] env[61906]: value = "task-1333609" [ 1034.193015] env[61906]: _type = "Task" [ 1034.193015] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.206028] env[61906]: DEBUG oslo_vmware.api [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333609, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.237563] env[61906]: DEBUG nova.network.neutron [req-7f160134-a2ce-4a99-a25e-f0f906be0bc5 req-874a5de8-e6bd-453a-8744-990517a431e4 service nova] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Updated VIF entry in instance network info cache for port e55c6ef7-fa51-455b-a1da-2fea0facd0a2. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1034.237937] env[61906]: DEBUG nova.network.neutron [req-7f160134-a2ce-4a99-a25e-f0f906be0bc5 req-874a5de8-e6bd-453a-8744-990517a431e4 service nova] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Updating instance_info_cache with network_info: [{"id": "e55c6ef7-fa51-455b-a1da-2fea0facd0a2", "address": "fa:16:3e:86:a7:d3", "network": {"id": "af039bdd-ffd2-483d-9655-e7df48936290", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-111531540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f7fc2042e6549cabe98838990f35842", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape55c6ef7-fa", "ovs_interfaceid": "e55c6ef7-fa51-455b-a1da-2fea0facd0a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.520992] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "refresh_cache-1be91e16-1acb-4f13-9b26-cae496a0efb8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.521208] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquired lock "refresh_cache-1be91e16-1acb-4f13-9b26-cae496a0efb8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.521397] env[61906]: DEBUG nova.network.neutron [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1034.566208] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]527ea359-99ea-df86-fed7-7db07be80760, 'name': SearchDatastore_Task, 'duration_secs': 0.01418} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.566590] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.566877] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1034.567164] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.567353] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.567632] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1034.568351] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-59f1805f-ae36-4d69-a3af-002d57b95d5f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.583446] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1034.583650] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1034.584581] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4b811d4-c0df-492e-a346-bd70c77bb374 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.591127] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 1034.591127] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5280a1db-06e3-28d4-eeb7-0a279feabc23" [ 1034.591127] env[61906]: _type = "Task" [ 1034.591127] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.600862] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5280a1db-06e3-28d4-eeb7-0a279feabc23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.648773] env[61906]: DEBUG nova.compute.manager [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1034.704076] env[61906]: DEBUG oslo_vmware.api [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333609, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.741431] env[61906]: DEBUG oslo_concurrency.lockutils [req-7f160134-a2ce-4a99-a25e-f0f906be0bc5 req-874a5de8-e6bd-453a-8744-990517a431e4 service nova] Releasing lock "refresh_cache-ed213a8b-7d7e-47e1-8a4f-d657c13df3ce" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.996260] env[61906]: DEBUG nova.compute.manager [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1034.997465] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e119900c-db07-492c-b475-9f77b8edac15 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.054014] env[61906]: DEBUG nova.network.neutron [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1035.103458] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5280a1db-06e3-28d4-eeb7-0a279feabc23, 'name': SearchDatastore_Task, 'duration_secs': 0.010188} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.104306] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8e9e27b-99c6-428d-95ca-8ccabfbfdc57 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.109824] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 1035.109824] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]525b3b7b-46c4-72c2-30a9-d158b35fc701" [ 1035.109824] env[61906]: _type = "Task" [ 1035.109824] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.117810] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]525b3b7b-46c4-72c2-30a9-d158b35fc701, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.174706] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.174976] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.176486] env[61906]: INFO nova.compute.claims [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1035.204632] env[61906]: DEBUG oslo_vmware.api [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333609, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.205496] env[61906]: DEBUG nova.network.neutron [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Updating instance_info_cache with network_info: [{"id": "df3f0533-5c71-429f-9e5b-a39a6eb52ff8", "address": "fa:16:3e:bc:61:4f", "network": {"id": "c3bcef0f-44a7-44f7-8130-df77c0d3ac2d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-330384233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "129cdb06b47346b6895df45e73f7b2dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35e463c7-7d78-4d66-8efd-6127b1f3ee17", "external-id": "nsx-vlan-transportzone-175", "segmentation_id": 175, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf3f0533-5c", "ovs_interfaceid": "df3f0533-5c71-429f-9e5b-a39a6eb52ff8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.294721] env[61906]: DEBUG nova.compute.manager [req-b8aff0e9-23ad-4e5f-a4da-1c512085d555 req-09029ef1-1d23-4e36-8bbc-fbc14761df31 service nova] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Received event network-changed-df3f0533-5c71-429f-9e5b-a39a6eb52ff8 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1035.295133] env[61906]: DEBUG nova.compute.manager [req-b8aff0e9-23ad-4e5f-a4da-1c512085d555 req-09029ef1-1d23-4e36-8bbc-fbc14761df31 service nova] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Refreshing instance network info cache due to event network-changed-df3f0533-5c71-429f-9e5b-a39a6eb52ff8. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1035.295557] env[61906]: DEBUG oslo_concurrency.lockutils [req-b8aff0e9-23ad-4e5f-a4da-1c512085d555 req-09029ef1-1d23-4e36-8bbc-fbc14761df31 service nova] Acquiring lock "refresh_cache-1be91e16-1acb-4f13-9b26-cae496a0efb8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.512228] env[61906]: INFO nova.compute.manager [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] instance snapshotting [ 1035.515709] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd121772-9643-4a72-a461-756720303d33 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.538261] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba156598-a1dc-45fe-84bf-6b01e52b4bce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.624215] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]525b3b7b-46c4-72c2-30a9-d158b35fc701, 'name': SearchDatastore_Task, 'duration_secs': 0.009759} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.624215] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.624215] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] ed213a8b-7d7e-47e1-8a4f-d657c13df3ce/ed213a8b-7d7e-47e1-8a4f-d657c13df3ce.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1035.624215] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e71d374b-73b7-41e5-8bb7-a063642046a5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.631027] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 1035.631027] env[61906]: value = "task-1333610" [ 1035.631027] env[61906]: _type = "Task" [ 1035.631027] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.639536] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333610, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.704549] env[61906]: DEBUG oslo_vmware.api [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333609, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.708138] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Releasing lock "refresh_cache-1be91e16-1acb-4f13-9b26-cae496a0efb8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.708448] env[61906]: DEBUG nova.compute.manager [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Instance network_info: |[{"id": "df3f0533-5c71-429f-9e5b-a39a6eb52ff8", "address": "fa:16:3e:bc:61:4f", "network": {"id": "c3bcef0f-44a7-44f7-8130-df77c0d3ac2d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-330384233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "129cdb06b47346b6895df45e73f7b2dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35e463c7-7d78-4d66-8efd-6127b1f3ee17", "external-id": "nsx-vlan-transportzone-175", "segmentation_id": 175, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf3f0533-5c", "ovs_interfaceid": "df3f0533-5c71-429f-9e5b-a39a6eb52ff8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1035.708776] env[61906]: DEBUG oslo_concurrency.lockutils [req-b8aff0e9-23ad-4e5f-a4da-1c512085d555 req-09029ef1-1d23-4e36-8bbc-fbc14761df31 service nova] Acquired lock "refresh_cache-1be91e16-1acb-4f13-9b26-cae496a0efb8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.708986] env[61906]: DEBUG nova.network.neutron [req-b8aff0e9-23ad-4e5f-a4da-1c512085d555 req-09029ef1-1d23-4e36-8bbc-fbc14761df31 service nova] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Refreshing network info cache for port df3f0533-5c71-429f-9e5b-a39a6eb52ff8 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1035.710245] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bc:61:4f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35e463c7-7d78-4d66-8efd-6127b1f3ee17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'df3f0533-5c71-429f-9e5b-a39a6eb52ff8', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1035.717718] env[61906]: DEBUG oslo.service.loopingcall [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1035.718561] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1035.718790] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-52ec333e-ab1f-48c4-ba35-37744f56b1b9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.739382] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1035.739382] env[61906]: value = "task-1333611" [ 1035.739382] env[61906]: _type = "Task" [ 1035.739382] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.747824] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333611, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.050925] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Creating Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1036.051456] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-49ea0a1a-a2d9-42c2-b678-5f6eb4b654d0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.060509] env[61906]: DEBUG oslo_vmware.api [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 1036.060509] env[61906]: value = "task-1333612" [ 1036.060509] env[61906]: _type = "Task" [ 1036.060509] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.071227] env[61906]: DEBUG oslo_vmware.api [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333612, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.143626] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333610, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.205646] env[61906]: DEBUG oslo_vmware.api [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333609, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.249888] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333611, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.328287] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25441c9c-f83d-454d-bf39-dd09a00379c6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.336135] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31efdcb6-38da-4277-b6bc-6f8847f499da {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.085122] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a06cd72-3b9b-4aae-852e-1129c505b2cb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.094175] env[61906]: DEBUG oslo_vmware.api [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333612, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.102694] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333610, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.588836} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.103090] env[61906]: DEBUG oslo_vmware.api [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333609, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.103586] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] ed213a8b-7d7e-47e1-8a4f-d657c13df3ce/ed213a8b-7d7e-47e1-8a4f-d657c13df3ce.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1037.103846] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1037.104419] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-41e83f0e-071a-449d-866e-21bbf2e68091 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.110864] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333611, 'name': CreateVM_Task, 'duration_secs': 0.574483} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.113378] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1037.118021] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b7f57b6-c569-49e1-b948-acf7e9dc9aee {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.119870] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.120054] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.120383] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1037.120992] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 1037.120992] env[61906]: value = "task-1333613" [ 1037.120992] env[61906]: _type = "Task" [ 1037.120992] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.121210] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a12b09f-7acd-46d6-b274-e18845a3e9d8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.133990] env[61906]: DEBUG nova.compute.provider_tree [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1037.137250] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1037.137250] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]525eac56-d4f3-6ea7-839c-1cf2776e8dd2" [ 1037.137250] env[61906]: _type = "Task" [ 1037.137250] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.143496] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333613, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.150604] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]525eac56-d4f3-6ea7-839c-1cf2776e8dd2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.221422] env[61906]: DEBUG nova.network.neutron [req-b8aff0e9-23ad-4e5f-a4da-1c512085d555 req-09029ef1-1d23-4e36-8bbc-fbc14761df31 service nova] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Updated VIF entry in instance network info cache for port df3f0533-5c71-429f-9e5b-a39a6eb52ff8. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1037.221806] env[61906]: DEBUG nova.network.neutron [req-b8aff0e9-23ad-4e5f-a4da-1c512085d555 req-09029ef1-1d23-4e36-8bbc-fbc14761df31 service nova] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Updating instance_info_cache with network_info: [{"id": "df3f0533-5c71-429f-9e5b-a39a6eb52ff8", "address": "fa:16:3e:bc:61:4f", "network": {"id": "c3bcef0f-44a7-44f7-8130-df77c0d3ac2d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-330384233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "129cdb06b47346b6895df45e73f7b2dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35e463c7-7d78-4d66-8efd-6127b1f3ee17", "external-id": "nsx-vlan-transportzone-175", "segmentation_id": 175, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf3f0533-5c", "ovs_interfaceid": "df3f0533-5c71-429f-9e5b-a39a6eb52ff8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.584650] env[61906]: DEBUG oslo_vmware.api [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333612, 'name': CreateSnapshot_Task, 'duration_secs': 1.373348} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.587296] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Created Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1037.587538] env[61906]: DEBUG oslo_vmware.api [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333609, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.588244] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae45c824-59a0-4e19-980f-cf6b405a214b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.632799] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333613, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067483} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.633872] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1037.634656] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4719acf1-744e-44d4-8e2a-f157951dcc9a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.637834] env[61906]: DEBUG nova.scheduler.client.report [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1037.661866] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] ed213a8b-7d7e-47e1-8a4f-d657c13df3ce/ed213a8b-7d7e-47e1-8a4f-d657c13df3ce.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1037.662657] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2112ce5-c0fb-453e-acb4-f770563b9438 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.680162] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]525eac56-d4f3-6ea7-839c-1cf2776e8dd2, 'name': SearchDatastore_Task, 'duration_secs': 0.01671} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.680772] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.681010] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1037.681252] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.681408] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.681583] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1037.681838] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3dc87036-6e42-4104-b7fc-d6cdadfdbff3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.686500] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 1037.686500] env[61906]: value = "task-1333614" [ 1037.686500] env[61906]: _type = "Task" [ 1037.686500] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.691688] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1037.691890] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1037.695387] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea19c4ef-da27-4572-8153-5817e2d8d339 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.697750] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333614, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.700767] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1037.700767] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]529b6f95-603b-388d-a23c-68b8a98edf77" [ 1037.700767] env[61906]: _type = "Task" [ 1037.700767] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.709122] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]529b6f95-603b-388d-a23c-68b8a98edf77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.724216] env[61906]: DEBUG oslo_concurrency.lockutils [req-b8aff0e9-23ad-4e5f-a4da-1c512085d555 req-09029ef1-1d23-4e36-8bbc-fbc14761df31 service nova] Releasing lock "refresh_cache-1be91e16-1acb-4f13-9b26-cae496a0efb8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.082603] env[61906]: DEBUG oslo_vmware.api [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333609, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.106038] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Creating linked-clone VM from snapshot {{(pid=61906) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1038.106547] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-fcc8fb4f-7c75-43ea-8ba3-64c7f005558c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.118459] env[61906]: DEBUG oslo_vmware.api [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 1038.118459] env[61906]: value = "task-1333615" [ 1038.118459] env[61906]: _type = "Task" [ 1038.118459] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.129444] env[61906]: DEBUG oslo_vmware.api [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333615, 'name': CloneVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.142584] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.967s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.143088] env[61906]: DEBUG nova.compute.manager [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1038.197598] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333614, 'name': ReconfigVM_Task, 'duration_secs': 0.346148} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.197921] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Reconfigured VM instance instance-00000061 to attach disk [datastore1] ed213a8b-7d7e-47e1-8a4f-d657c13df3ce/ed213a8b-7d7e-47e1-8a4f-d657c13df3ce.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1038.198621] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-017c5382-db04-4a8c-9fae-a46462793536 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.207105] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 1038.207105] env[61906]: value = "task-1333616" [ 1038.207105] env[61906]: _type = "Task" [ 1038.207105] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.210982] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]529b6f95-603b-388d-a23c-68b8a98edf77, 'name': SearchDatastore_Task, 'duration_secs': 0.009111} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.214662] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc8e66dc-6c9d-49b8-84dc-248d44ea8523 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.223485] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1038.223485] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5203417d-f487-cbee-0b40-7ce8ae664283" [ 1038.223485] env[61906]: _type = "Task" [ 1038.223485] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.226972] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333616, 'name': Rename_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.237584] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5203417d-f487-cbee-0b40-7ce8ae664283, 'name': SearchDatastore_Task, 'duration_secs': 0.011767} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.237885] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.238165] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 1be91e16-1acb-4f13-9b26-cae496a0efb8/1be91e16-1acb-4f13-9b26-cae496a0efb8.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1038.238779] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-10ddf2e7-bb5e-4189-98f4-9785cd96e7c7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.257023] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1038.257023] env[61906]: value = "task-1333617" [ 1038.257023] env[61906]: _type = "Task" [ 1038.257023] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.266818] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333617, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.585396] env[61906]: DEBUG oslo_vmware.api [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333609, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.630309] env[61906]: DEBUG oslo_vmware.api [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333615, 'name': CloneVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.649262] env[61906]: DEBUG nova.compute.utils [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1038.651277] env[61906]: DEBUG nova.compute.manager [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1038.651505] env[61906]: DEBUG nova.network.neutron [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1038.699601] env[61906]: DEBUG nova.policy [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b49e5a6aeb1d43e290a645e14861b889', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '53644a4143a24759a9ff2b5e28b84fb4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1038.722297] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333616, 'name': Rename_Task, 'duration_secs': 0.19425} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.722438] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1038.722660] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8aae9715-16f5-479e-b01a-7f517b049931 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.732090] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 1038.732090] env[61906]: value = "task-1333618" [ 1038.732090] env[61906]: _type = "Task" [ 1038.732090] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.743703] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333618, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.771175] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333617, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.091273] env[61906]: DEBUG oslo_vmware.api [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333609, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.131868] env[61906]: DEBUG oslo_vmware.api [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333615, 'name': CloneVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.154782] env[61906]: DEBUG nova.compute.manager [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1039.172285] env[61906]: DEBUG nova.network.neutron [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Successfully created port: e2c74014-2d68-4807-86d9-3009eee2db02 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1039.245576] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333618, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.273795] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333617, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.619386} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.274064] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 1be91e16-1acb-4f13-9b26-cae496a0efb8/1be91e16-1acb-4f13-9b26-cae496a0efb8.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1039.274291] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1039.274563] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f878c0a1-32a0-4dfd-a3ca-72bc4591b9d8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.284227] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1039.284227] env[61906]: value = "task-1333619" [ 1039.284227] env[61906]: _type = "Task" [ 1039.284227] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.293932] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333619, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.342494] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "86be7197-c0d1-49c5-aa30-cf908a506031" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.342791] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "86be7197-c0d1-49c5-aa30-cf908a506031" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.586145] env[61906]: DEBUG oslo_vmware.api [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333609, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.632179] env[61906]: DEBUG oslo_vmware.api [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333615, 'name': CloneVM_Task} progress is 95%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.743339] env[61906]: DEBUG oslo_vmware.api [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333618, 'name': PowerOnVM_Task, 'duration_secs': 0.591696} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.743644] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1039.743850] env[61906]: INFO nova.compute.manager [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Took 8.89 seconds to spawn the instance on the hypervisor. [ 1039.744039] env[61906]: DEBUG nova.compute.manager [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1039.744938] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be7695ee-185e-4555-85f6-0f3c0b1ae092 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.797125] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333619, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.130951} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.797356] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1039.798158] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f89d93-5e8b-4f8b-9e45-78d889e90158 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.820260] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] 1be91e16-1acb-4f13-9b26-cae496a0efb8/1be91e16-1acb-4f13-9b26-cae496a0efb8.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1039.820598] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2cf5ff5c-6f64-4758-8a2a-951a17866e02 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.841084] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.841336] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.846021] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1039.846021] env[61906]: value = "task-1333620" [ 1039.846021] env[61906]: _type = "Task" [ 1039.846021] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.846297] env[61906]: DEBUG nova.compute.manager [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1039.855129] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333620, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.087265] env[61906]: DEBUG oslo_vmware.api [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333609, 'name': ReconfigVM_Task, 'duration_secs': 5.804654} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.087265] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Releasing lock "7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.087265] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Reconfigured VM to detach interface {{(pid=61906) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1040.133829] env[61906]: DEBUG oslo_vmware.api [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333615, 'name': CloneVM_Task, 'duration_secs': 1.669483} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.134191] env[61906]: INFO nova.virt.vmwareapi.vmops [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Created linked-clone VM from snapshot [ 1040.135023] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b96b4630-6813-4b31-8e9d-29588940e54f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.147111] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Uploading image 539c0530-b53d-4eb6-9e5d-8f1523a03f18 {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1040.163026] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Destroying the VM {{(pid=61906) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1040.163026] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f0c53c96-5bbd-45c2-93a1-907766a7ed9a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.168047] env[61906]: DEBUG nova.compute.manager [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1040.174045] env[61906]: DEBUG oslo_vmware.api [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 1040.174045] env[61906]: value = "task-1333621" [ 1040.174045] env[61906]: _type = "Task" [ 1040.174045] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.184239] env[61906]: DEBUG oslo_vmware.api [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333621, 'name': Destroy_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.203090] env[61906]: DEBUG nova.virt.hardware [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1040.203090] env[61906]: DEBUG nova.virt.hardware [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1040.203090] env[61906]: DEBUG nova.virt.hardware [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1040.203090] env[61906]: DEBUG nova.virt.hardware [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1040.203090] env[61906]: DEBUG nova.virt.hardware [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1040.203090] env[61906]: DEBUG nova.virt.hardware [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1040.205391] env[61906]: DEBUG nova.virt.hardware [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1040.206145] env[61906]: DEBUG nova.virt.hardware [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1040.206561] env[61906]: DEBUG nova.virt.hardware [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1040.206890] env[61906]: DEBUG nova.virt.hardware [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1040.207296] env[61906]: DEBUG nova.virt.hardware [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1040.208560] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f7c3a45-8fb4-4998-82ea-4cd142c000e0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.217596] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfea9edd-a779-4d87-bbb4-165113209d44 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.267679] env[61906]: INFO nova.compute.manager [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Took 18.62 seconds to build instance. [ 1040.343333] env[61906]: DEBUG nova.compute.manager [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1040.357777] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333620, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.380525] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.380842] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.382427] env[61906]: INFO nova.compute.claims [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1040.548835] env[61906]: DEBUG oslo_concurrency.lockutils [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "b1159533-c970-49d6-af42-b954b20d92fb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.549165] env[61906]: DEBUG oslo_concurrency.lockutils [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "b1159533-c970-49d6-af42-b954b20d92fb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.549396] env[61906]: DEBUG oslo_concurrency.lockutils [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "b1159533-c970-49d6-af42-b954b20d92fb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.549625] env[61906]: DEBUG oslo_concurrency.lockutils [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "b1159533-c970-49d6-af42-b954b20d92fb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.549940] env[61906]: DEBUG oslo_concurrency.lockutils [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "b1159533-c970-49d6-af42-b954b20d92fb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.552996] env[61906]: INFO nova.compute.manager [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Terminating instance [ 1040.556334] env[61906]: DEBUG nova.compute.manager [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1040.556552] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1040.557703] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b631d307-4799-430d-a29d-9cbe2ea1c6e3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.565727] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1040.566121] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0ec8d79a-d74e-4337-9f56-12693a92c0cd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.572635] env[61906]: DEBUG oslo_vmware.api [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 1040.572635] env[61906]: value = "task-1333622" [ 1040.572635] env[61906]: _type = "Task" [ 1040.572635] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.582794] env[61906]: DEBUG oslo_vmware.api [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333622, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.683182] env[61906]: DEBUG oslo_vmware.api [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333621, 'name': Destroy_Task, 'duration_secs': 0.373701} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.683182] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Destroyed the VM [ 1040.683182] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Deleting Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1040.683527] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-6ef14718-c75e-46b9-b5f4-398d41d35965 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.692460] env[61906]: DEBUG oslo_vmware.api [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 1040.692460] env[61906]: value = "task-1333623" [ 1040.692460] env[61906]: _type = "Task" [ 1040.692460] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.701144] env[61906]: DEBUG oslo_vmware.api [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333623, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.771579] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6da5dc81-0031-4eb6-ab86-92458a5e33f4 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "ed213a8b-7d7e-47e1-8a4f-d657c13df3ce" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.135s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.777817] env[61906]: DEBUG nova.compute.manager [req-7c9f96f6-2d7c-4edb-8418-940ecfb02c2f req-48aae99d-6784-4612-9ace-195b2efa9693 service nova] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Received event network-vif-plugged-e2c74014-2d68-4807-86d9-3009eee2db02 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1040.778079] env[61906]: DEBUG oslo_concurrency.lockutils [req-7c9f96f6-2d7c-4edb-8418-940ecfb02c2f req-48aae99d-6784-4612-9ace-195b2efa9693 service nova] Acquiring lock "36c6b1a3-be22-4cec-b21c-0445f5988be1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.778570] env[61906]: DEBUG oslo_concurrency.lockutils [req-7c9f96f6-2d7c-4edb-8418-940ecfb02c2f req-48aae99d-6784-4612-9ace-195b2efa9693 service nova] Lock "36c6b1a3-be22-4cec-b21c-0445f5988be1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.778667] env[61906]: DEBUG oslo_concurrency.lockutils [req-7c9f96f6-2d7c-4edb-8418-940ecfb02c2f req-48aae99d-6784-4612-9ace-195b2efa9693 service nova] Lock "36c6b1a3-be22-4cec-b21c-0445f5988be1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.778852] env[61906]: DEBUG nova.compute.manager [req-7c9f96f6-2d7c-4edb-8418-940ecfb02c2f req-48aae99d-6784-4612-9ace-195b2efa9693 service nova] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] No waiting events found dispatching network-vif-plugged-e2c74014-2d68-4807-86d9-3009eee2db02 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1040.779030] env[61906]: WARNING nova.compute.manager [req-7c9f96f6-2d7c-4edb-8418-940ecfb02c2f req-48aae99d-6784-4612-9ace-195b2efa9693 service nova] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Received unexpected event network-vif-plugged-e2c74014-2d68-4807-86d9-3009eee2db02 for instance with vm_state building and task_state spawning. [ 1040.858054] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333620, 'name': ReconfigVM_Task, 'duration_secs': 0.55303} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.858402] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Reconfigured VM instance instance-00000062 to attach disk [datastore1] 1be91e16-1acb-4f13-9b26-cae496a0efb8/1be91e16-1acb-4f13-9b26-cae496a0efb8.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1040.859117] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d6abb82a-ea87-445a-b6ce-051808dd9a7b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.865585] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1040.865585] env[61906]: value = "task-1333624" [ 1040.865585] env[61906]: _type = "Task" [ 1040.865585] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.872712] env[61906]: DEBUG nova.network.neutron [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Successfully updated port: e2c74014-2d68-4807-86d9-3009eee2db02 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1040.878454] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333624, 'name': Rename_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.879509] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.085195] env[61906]: DEBUG oslo_vmware.api [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333622, 'name': PowerOffVM_Task, 'duration_secs': 0.250108} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.085578] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1041.085751] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1041.086049] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3ee569c6-3476-40bd-b5f1-d30b0bb34765 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.200449] env[61906]: DEBUG oslo_vmware.api [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333623, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.276208] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1041.276483] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1041.276909] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Deleting the datastore file [datastore2] b1159533-c970-49d6-af42-b954b20d92fb {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1041.276909] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-db88a14e-aa8f-4227-95ba-dfef3ab90e00 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.283950] env[61906]: DEBUG oslo_vmware.api [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 1041.283950] env[61906]: value = "task-1333626" [ 1041.283950] env[61906]: _type = "Task" [ 1041.283950] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.291363] env[61906]: DEBUG oslo_vmware.api [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333626, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.376088] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "refresh_cache-36c6b1a3-be22-4cec-b21c-0445f5988be1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.376253] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired lock "refresh_cache-36c6b1a3-be22-4cec-b21c-0445f5988be1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.376408] env[61906]: DEBUG nova.network.neutron [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1041.377553] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333624, 'name': Rename_Task, 'duration_secs': 0.176013} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.378155] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1041.378285] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-26483f7c-f9f6-465a-9912-da8519ecd535 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.385302] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1041.385302] env[61906]: value = "task-1333627" [ 1041.385302] env[61906]: _type = "Task" [ 1041.385302] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.396355] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333627, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.522223] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a44ab9c-a907-40e8-8585-60fdb6ee8d47 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.529859] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91489e77-d28a-410c-a855-76a667b777b3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.563543] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b93703a3-9676-4e25-b5cb-23f7fb8fc44d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.571917] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4313dc8d-fada-4a46-94c0-9b3cd650e263 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.585705] env[61906]: DEBUG nova.compute.provider_tree [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.649637] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.649842] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.650044] env[61906]: DEBUG nova.network.neutron [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1041.681354] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "ed213a8b-7d7e-47e1-8a4f-d657c13df3ce" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.681651] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "ed213a8b-7d7e-47e1-8a4f-d657c13df3ce" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.681867] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "ed213a8b-7d7e-47e1-8a4f-d657c13df3ce-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.682083] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "ed213a8b-7d7e-47e1-8a4f-d657c13df3ce-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.682269] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "ed213a8b-7d7e-47e1-8a4f-d657c13df3ce-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.684471] env[61906]: INFO nova.compute.manager [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Terminating instance [ 1041.686173] env[61906]: DEBUG nova.compute.manager [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1041.686375] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1041.687197] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f734d4c-32e1-49d5-80fc-1d27236b0b34 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.696895] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1041.697461] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6fc98e66-1439-46e3-bda6-774778ffe129 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.702212] env[61906]: DEBUG oslo_vmware.api [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333623, 'name': RemoveSnapshot_Task, 'duration_secs': 0.81796} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.703289] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Deleted Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1041.705550] env[61906]: DEBUG oslo_vmware.api [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 1041.705550] env[61906]: value = "task-1333628" [ 1041.705550] env[61906]: _type = "Task" [ 1041.705550] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.713140] env[61906]: DEBUG oslo_vmware.api [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333628, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.795927] env[61906]: DEBUG oslo_vmware.api [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333626, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.242433} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.796261] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1041.796577] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1041.796874] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1041.797115] env[61906]: INFO nova.compute.manager [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Took 1.24 seconds to destroy the instance on the hypervisor. [ 1041.797378] env[61906]: DEBUG oslo.service.loopingcall [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1041.797587] env[61906]: DEBUG nova.compute.manager [-] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1041.797714] env[61906]: DEBUG nova.network.neutron [-] [instance: b1159533-c970-49d6-af42-b954b20d92fb] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1041.896814] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333627, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.921044] env[61906]: DEBUG nova.network.neutron [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1042.088669] env[61906]: DEBUG nova.scheduler.client.report [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1042.209033] env[61906]: WARNING nova.compute.manager [None req-e4b4c32c-dbf9-4e65-880f-040306aeaf51 tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Image not found during snapshot: nova.exception.ImageNotFound: Image 539c0530-b53d-4eb6-9e5d-8f1523a03f18 could not be found. [ 1042.211267] env[61906]: DEBUG nova.network.neutron [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Updating instance_info_cache with network_info: [{"id": "e2c74014-2d68-4807-86d9-3009eee2db02", "address": "fa:16:3e:82:25:bf", "network": {"id": "bc9185e9-f26b-489c-84ee-e6db41ac6283", "bridge": "br-int", "label": "tempest-ServersTestJSON-1167747148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53644a4143a24759a9ff2b5e28b84fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape2c74014-2d", "ovs_interfaceid": "e2c74014-2d68-4807-86d9-3009eee2db02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.225134] env[61906]: DEBUG oslo_vmware.api [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333628, 'name': PowerOffVM_Task, 'duration_secs': 0.18643} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.226235] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1042.226472] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1042.227094] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f01dd188-5512-4474-81d8-bd3adbac5980 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.292431] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1042.292705] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1042.292898] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Deleting the datastore file [datastore1] ed213a8b-7d7e-47e1-8a4f-d657c13df3ce {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1042.293245] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-893f22a5-8b91-42c6-8d3d-e9c4bdd43653 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.299756] env[61906]: DEBUG oslo_vmware.api [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for the task: (returnval){ [ 1042.299756] env[61906]: value = "task-1333630" [ 1042.299756] env[61906]: _type = "Task" [ 1042.299756] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.309536] env[61906]: DEBUG oslo_vmware.api [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333630, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.398400] env[61906]: DEBUG oslo_vmware.api [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333627, 'name': PowerOnVM_Task, 'duration_secs': 0.682647} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.398688] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1042.398887] env[61906]: INFO nova.compute.manager [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Took 9.32 seconds to spawn the instance on the hypervisor. [ 1042.399105] env[61906]: DEBUG nova.compute.manager [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1042.399882] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-905453a4-a604-4a5e-90c7-71681f177752 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.571670] env[61906]: INFO nova.network.neutron [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Port a701c10b-b79f-4f4e-a4e0-7d702211ffab from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1042.571670] env[61906]: DEBUG nova.network.neutron [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Updating instance_info_cache with network_info: [{"id": "73d0e617-fd5e-4fa2-81c5-710a3beac1a4", "address": "fa:16:3e:66:bf:d9", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73d0e617-fd", "ovs_interfaceid": "73d0e617-fd5e-4fa2-81c5-710a3beac1a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.594122] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.213s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.594710] env[61906]: DEBUG nova.compute.manager [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1042.598156] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.719s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.603721] env[61906]: INFO nova.compute.claims [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1042.714584] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Releasing lock "refresh_cache-36c6b1a3-be22-4cec-b21c-0445f5988be1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.714923] env[61906]: DEBUG nova.compute.manager [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Instance network_info: |[{"id": "e2c74014-2d68-4807-86d9-3009eee2db02", "address": "fa:16:3e:82:25:bf", "network": {"id": "bc9185e9-f26b-489c-84ee-e6db41ac6283", "bridge": "br-int", "label": "tempest-ServersTestJSON-1167747148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53644a4143a24759a9ff2b5e28b84fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape2c74014-2d", "ovs_interfaceid": "e2c74014-2d68-4807-86d9-3009eee2db02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1042.715597] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:82:25:bf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e2c74014-2d68-4807-86d9-3009eee2db02', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1042.724815] env[61906]: DEBUG oslo.service.loopingcall [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1042.725469] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1042.727786] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4b6d90a0-ab21-4e83-affc-68615e8a424f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.748448] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1042.748448] env[61906]: value = "task-1333631" [ 1042.748448] env[61906]: _type = "Task" [ 1042.748448] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.756049] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333631, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.792030] env[61906]: DEBUG nova.compute.manager [req-a2677b6c-5e83-439b-a29c-2b5370d391f3 req-435046c6-3f1c-4d0c-9bd2-6736ff0d1346 service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Received event network-changed-73d0e617-fd5e-4fa2-81c5-710a3beac1a4 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1042.793270] env[61906]: DEBUG nova.compute.manager [req-a2677b6c-5e83-439b-a29c-2b5370d391f3 req-435046c6-3f1c-4d0c-9bd2-6736ff0d1346 service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Refreshing instance network info cache due to event network-changed-73d0e617-fd5e-4fa2-81c5-710a3beac1a4. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1042.793270] env[61906]: DEBUG oslo_concurrency.lockutils [req-a2677b6c-5e83-439b-a29c-2b5370d391f3 req-435046c6-3f1c-4d0c-9bd2-6736ff0d1346 service nova] Acquiring lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.813153] env[61906]: DEBUG oslo_vmware.api [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Task: {'id': task-1333630, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.410598} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.813485] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1042.813696] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1042.814011] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1042.814126] env[61906]: INFO nova.compute.manager [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1042.814374] env[61906]: DEBUG oslo.service.loopingcall [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1042.814603] env[61906]: DEBUG nova.compute.manager [-] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1042.814715] env[61906]: DEBUG nova.network.neutron [-] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1042.888406] env[61906]: DEBUG nova.compute.manager [req-afe2a213-a0eb-42e2-b929-3495ab6d1663 req-10704f12-a244-4d53-b727-b0a62a7cdca5 service nova] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Received event network-changed-e2c74014-2d68-4807-86d9-3009eee2db02 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1042.888574] env[61906]: DEBUG nova.compute.manager [req-afe2a213-a0eb-42e2-b929-3495ab6d1663 req-10704f12-a244-4d53-b727-b0a62a7cdca5 service nova] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Refreshing instance network info cache due to event network-changed-e2c74014-2d68-4807-86d9-3009eee2db02. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1042.888786] env[61906]: DEBUG oslo_concurrency.lockutils [req-afe2a213-a0eb-42e2-b929-3495ab6d1663 req-10704f12-a244-4d53-b727-b0a62a7cdca5 service nova] Acquiring lock "refresh_cache-36c6b1a3-be22-4cec-b21c-0445f5988be1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.888926] env[61906]: DEBUG oslo_concurrency.lockutils [req-afe2a213-a0eb-42e2-b929-3495ab6d1663 req-10704f12-a244-4d53-b727-b0a62a7cdca5 service nova] Acquired lock "refresh_cache-36c6b1a3-be22-4cec-b21c-0445f5988be1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.889465] env[61906]: DEBUG nova.network.neutron [req-afe2a213-a0eb-42e2-b929-3495ab6d1663 req-10704f12-a244-4d53-b727-b0a62a7cdca5 service nova] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Refreshing network info cache for port e2c74014-2d68-4807-86d9-3009eee2db02 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1042.924911] env[61906]: INFO nova.compute.manager [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Took 18.66 seconds to build instance. [ 1043.076218] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Releasing lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.078347] env[61906]: DEBUG oslo_concurrency.lockutils [req-a2677b6c-5e83-439b-a29c-2b5370d391f3 req-435046c6-3f1c-4d0c-9bd2-6736ff0d1346 service nova] Acquired lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.079431] env[61906]: DEBUG nova.network.neutron [req-a2677b6c-5e83-439b-a29c-2b5370d391f3 req-435046c6-3f1c-4d0c-9bd2-6736ff0d1346 service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Refreshing network info cache for port 73d0e617-fd5e-4fa2-81c5-710a3beac1a4 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1043.109199] env[61906]: DEBUG nova.compute.utils [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1043.118339] env[61906]: DEBUG nova.compute.manager [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1043.118729] env[61906]: DEBUG nova.network.neutron [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1043.160958] env[61906]: DEBUG nova.policy [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '315d6310be014ebdb4b868fd93a680a0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1d6fb2749df6498f82aaf7c5529b309f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1043.184225] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "0fe121ad-a539-4c04-bb65-b524cb3d91a8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.184225] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "0fe121ad-a539-4c04-bb65-b524cb3d91a8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.184225] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "0fe121ad-a539-4c04-bb65-b524cb3d91a8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.184225] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "0fe121ad-a539-4c04-bb65-b524cb3d91a8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.184225] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "0fe121ad-a539-4c04-bb65-b524cb3d91a8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.186894] env[61906]: INFO nova.compute.manager [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Terminating instance [ 1043.189087] env[61906]: DEBUG nova.compute.manager [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1043.189399] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1043.190681] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c92c8605-7867-4e7f-b86b-6712cec2a723 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.198882] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1043.202311] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-93238d12-1cc7-4c1c-8bb1-d76e4bd85754 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.205664] env[61906]: DEBUG nova.network.neutron [-] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.208111] env[61906]: DEBUG oslo_vmware.api [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 1043.208111] env[61906]: value = "task-1333632" [ 1043.208111] env[61906]: _type = "Task" [ 1043.208111] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.217742] env[61906]: DEBUG oslo_vmware.api [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333632, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.259327] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333631, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.400967] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "interface-b3e9ab00-8627-4a57-839e-68be5f794b28-a701c10b-b79f-4f4e-a4e0-7d702211ffab" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.401278] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "interface-b3e9ab00-8627-4a57-839e-68be5f794b28-a701c10b-b79f-4f4e-a4e0-7d702211ffab" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.401645] env[61906]: DEBUG nova.objects.instance [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lazy-loading 'flavor' on Instance uuid b3e9ab00-8627-4a57-839e-68be5f794b28 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1043.416215] env[61906]: DEBUG nova.network.neutron [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Successfully created port: 24ab450e-e9c4-4d9a-9c4d-13dee30eb56b {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1043.427162] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82fa50bc-c40c-45d3-90de-3b27b62a1db2 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "1be91e16-1acb-4f13-9b26-cae496a0efb8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.171s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.581873] env[61906]: DEBUG nova.network.neutron [-] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.584102] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1ff2e62-d7e3-48ff-9213-d20adadac3f0 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "interface-7264af54-cd10-4507-8ae3-5bb8a5cd127f-a701c10b-b79f-4f4e-a4e0-7d702211ffab" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.968s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.625209] env[61906]: DEBUG nova.compute.manager [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1043.710466] env[61906]: INFO nova.compute.manager [-] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Took 1.91 seconds to deallocate network for instance. [ 1043.730905] env[61906]: DEBUG oslo_vmware.api [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333632, 'name': PowerOffVM_Task, 'duration_secs': 0.400892} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.731212] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1043.731489] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1043.731760] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6c991e4c-a31d-43c7-88a7-ae0fd434cc5d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.763744] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333631, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.803696] env[61906]: DEBUG nova.network.neutron [req-afe2a213-a0eb-42e2-b929-3495ab6d1663 req-10704f12-a244-4d53-b727-b0a62a7cdca5 service nova] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Updated VIF entry in instance network info cache for port e2c74014-2d68-4807-86d9-3009eee2db02. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1043.804084] env[61906]: DEBUG nova.network.neutron [req-afe2a213-a0eb-42e2-b929-3495ab6d1663 req-10704f12-a244-4d53-b727-b0a62a7cdca5 service nova] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Updating instance_info_cache with network_info: [{"id": "e2c74014-2d68-4807-86d9-3009eee2db02", "address": "fa:16:3e:82:25:bf", "network": {"id": "bc9185e9-f26b-489c-84ee-e6db41ac6283", "bridge": "br-int", "label": "tempest-ServersTestJSON-1167747148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53644a4143a24759a9ff2b5e28b84fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape2c74014-2d", "ovs_interfaceid": "e2c74014-2d68-4807-86d9-3009eee2db02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.819630] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1043.819713] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1043.820073] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Deleting the datastore file [datastore2] 0fe121ad-a539-4c04-bb65-b524cb3d91a8 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1043.823332] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-480527e4-2516-4587-9dbe-e4e622f10cce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.826971] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25ef5feb-cf83-4f8a-8a5e-87b163aff18f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.835662] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bb4aa56-912c-40a8-98b8-ad2c22669237 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.840594] env[61906]: DEBUG oslo_vmware.api [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for the task: (returnval){ [ 1043.840594] env[61906]: value = "task-1333634" [ 1043.840594] env[61906]: _type = "Task" [ 1043.840594] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.872545] env[61906]: DEBUG nova.network.neutron [req-a2677b6c-5e83-439b-a29c-2b5370d391f3 req-435046c6-3f1c-4d0c-9bd2-6736ff0d1346 service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Updated VIF entry in instance network info cache for port 73d0e617-fd5e-4fa2-81c5-710a3beac1a4. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1043.872926] env[61906]: DEBUG nova.network.neutron [req-a2677b6c-5e83-439b-a29c-2b5370d391f3 req-435046c6-3f1c-4d0c-9bd2-6736ff0d1346 service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Updating instance_info_cache with network_info: [{"id": "73d0e617-fd5e-4fa2-81c5-710a3beac1a4", "address": "fa:16:3e:66:bf:d9", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73d0e617-fd", "ovs_interfaceid": "73d0e617-fd5e-4fa2-81c5-710a3beac1a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.877216] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1c5073a-3236-4ed8-ac41-a0c61253d663 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.880429] env[61906]: DEBUG oslo_vmware.api [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333634, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.885623] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4583cdda-89a1-4c13-884e-f103c8befd82 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.900658] env[61906]: DEBUG nova.compute.provider_tree [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1044.023787] env[61906]: DEBUG nova.objects.instance [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lazy-loading 'pci_requests' on Instance uuid b3e9ab00-8627-4a57-839e-68be5f794b28 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.088207] env[61906]: INFO nova.compute.manager [-] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Took 1.27 seconds to deallocate network for instance. [ 1044.227598] env[61906]: DEBUG oslo_concurrency.lockutils [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.259669] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333631, 'name': CreateVM_Task, 'duration_secs': 1.409641} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.259812] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1044.260632] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.260842] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.261231] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1044.261556] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ebcda230-50d4-4922-bd0d-33e99bf2a4a6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.266088] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 1044.266088] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52e45f98-4746-ab4a-27de-b60b9f6c2301" [ 1044.266088] env[61906]: _type = "Task" [ 1044.266088] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.274664] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52e45f98-4746-ab4a-27de-b60b9f6c2301, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.307124] env[61906]: DEBUG oslo_concurrency.lockutils [req-afe2a213-a0eb-42e2-b929-3495ab6d1663 req-10704f12-a244-4d53-b727-b0a62a7cdca5 service nova] Releasing lock "refresh_cache-36c6b1a3-be22-4cec-b21c-0445f5988be1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.307426] env[61906]: DEBUG nova.compute.manager [req-afe2a213-a0eb-42e2-b929-3495ab6d1663 req-10704f12-a244-4d53-b727-b0a62a7cdca5 service nova] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Received event network-vif-deleted-ba5a12df-7917-4b2d-ac9e-7ef1380dabf7 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1044.307630] env[61906]: INFO nova.compute.manager [req-afe2a213-a0eb-42e2-b929-3495ab6d1663 req-10704f12-a244-4d53-b727-b0a62a7cdca5 service nova] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Neutron deleted interface ba5a12df-7917-4b2d-ac9e-7ef1380dabf7; detaching it from the instance and deleting it from the info cache [ 1044.307802] env[61906]: DEBUG nova.network.neutron [req-afe2a213-a0eb-42e2-b929-3495ab6d1663 req-10704f12-a244-4d53-b727-b0a62a7cdca5 service nova] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.348723] env[61906]: DEBUG oslo_vmware.api [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Task: {'id': task-1333634, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.460876} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.349012] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1044.349232] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1044.349416] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1044.349604] env[61906]: INFO nova.compute.manager [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1044.349923] env[61906]: DEBUG oslo.service.loopingcall [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1044.350145] env[61906]: DEBUG nova.compute.manager [-] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1044.350247] env[61906]: DEBUG nova.network.neutron [-] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1044.379235] env[61906]: DEBUG oslo_concurrency.lockutils [req-a2677b6c-5e83-439b-a29c-2b5370d391f3 req-435046c6-3f1c-4d0c-9bd2-6736ff0d1346 service nova] Releasing lock "refresh_cache-7264af54-cd10-4507-8ae3-5bb8a5cd127f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.403863] env[61906]: DEBUG nova.scheduler.client.report [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1044.526571] env[61906]: DEBUG nova.objects.base [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1044.526921] env[61906]: DEBUG nova.network.neutron [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1044.596131] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.608682] env[61906]: DEBUG nova.policy [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0de9abfd085f4a43a878da8448615bc4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e36ff7a500ef444cbf5a168c5a48208b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1044.636424] env[61906]: DEBUG nova.compute.manager [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1044.664015] env[61906]: DEBUG nova.virt.hardware [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1044.664389] env[61906]: DEBUG nova.virt.hardware [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1044.664578] env[61906]: DEBUG nova.virt.hardware [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1044.664796] env[61906]: DEBUG nova.virt.hardware [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1044.664965] env[61906]: DEBUG nova.virt.hardware [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1044.665143] env[61906]: DEBUG nova.virt.hardware [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1044.665375] env[61906]: DEBUG nova.virt.hardware [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1044.665548] env[61906]: DEBUG nova.virt.hardware [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1044.665742] env[61906]: DEBUG nova.virt.hardware [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1044.665916] env[61906]: DEBUG nova.virt.hardware [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1044.666114] env[61906]: DEBUG nova.virt.hardware [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1044.667042] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d63df96d-dd8e-4d37-932b-d5386b502ccc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.675805] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57ba752a-9909-46ee-ac31-be8434ea5ff8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.777030] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52e45f98-4746-ab4a-27de-b60b9f6c2301, 'name': SearchDatastore_Task, 'duration_secs': 0.010903} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.777920] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.777920] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1044.777920] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.778087] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.778122] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1044.778425] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-42a3cc4e-482d-488c-a424-d133726ed94d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.787108] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1044.787336] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1044.788137] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26e6f178-d047-4af6-8ee0-b7585f58a3cf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.793558] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 1044.793558] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5232b0de-930c-3389-9c81-7ee60567f040" [ 1044.793558] env[61906]: _type = "Task" [ 1044.793558] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.801785] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5232b0de-930c-3389-9c81-7ee60567f040, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.810620] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d69a7508-2968-4399-8d45-0fa14d0436e3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.819484] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb19bd72-c5ef-447e-a118-5a0cb3b5e338 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.845769] env[61906]: DEBUG nova.compute.manager [req-afe2a213-a0eb-42e2-b929-3495ab6d1663 req-10704f12-a244-4d53-b727-b0a62a7cdca5 service nova] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Detach interface failed, port_id=ba5a12df-7917-4b2d-ac9e-7ef1380dabf7, reason: Instance b1159533-c970-49d6-af42-b954b20d92fb could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1044.861011] env[61906]: DEBUG nova.compute.manager [req-0cc21e14-0ea5-404b-96ab-0f6a6cafd647 req-a40c5250-4dc9-4074-8f6d-eab55eda93f8 service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Received event network-changed-e0244018-d931-4a96-a25e-228b02a88f56 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1044.861163] env[61906]: DEBUG nova.compute.manager [req-0cc21e14-0ea5-404b-96ab-0f6a6cafd647 req-a40c5250-4dc9-4074-8f6d-eab55eda93f8 service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Refreshing instance network info cache due to event network-changed-e0244018-d931-4a96-a25e-228b02a88f56. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1044.861345] env[61906]: DEBUG oslo_concurrency.lockutils [req-0cc21e14-0ea5-404b-96ab-0f6a6cafd647 req-a40c5250-4dc9-4074-8f6d-eab55eda93f8 service nova] Acquiring lock "refresh_cache-b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.861500] env[61906]: DEBUG oslo_concurrency.lockutils [req-0cc21e14-0ea5-404b-96ab-0f6a6cafd647 req-a40c5250-4dc9-4074-8f6d-eab55eda93f8 service nova] Acquired lock "refresh_cache-b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.861669] env[61906]: DEBUG nova.network.neutron [req-0cc21e14-0ea5-404b-96ab-0f6a6cafd647 req-a40c5250-4dc9-4074-8f6d-eab55eda93f8 service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Refreshing network info cache for port e0244018-d931-4a96-a25e-228b02a88f56 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1044.909894] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.310s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.909894] env[61906]: DEBUG nova.compute.manager [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1044.911586] env[61906]: DEBUG oslo_concurrency.lockutils [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.684s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.911815] env[61906]: DEBUG nova.objects.instance [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lazy-loading 'resources' on Instance uuid b1159533-c970-49d6-af42-b954b20d92fb {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.924607] env[61906]: DEBUG nova.compute.manager [req-f8f8c64f-6b19-46a8-990b-de4614db2302 req-6755c298-2aae-4d1c-9bc3-7d2a80be733a service nova] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Received event network-vif-deleted-e55c6ef7-fa51-455b-a1da-2fea0facd0a2 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1044.924848] env[61906]: DEBUG nova.compute.manager [req-f8f8c64f-6b19-46a8-990b-de4614db2302 req-6755c298-2aae-4d1c-9bc3-7d2a80be733a service nova] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Received event network-vif-deleted-15a5be02-4db5-4752-ace0-a2e1bcec23c7 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1044.925044] env[61906]: INFO nova.compute.manager [req-f8f8c64f-6b19-46a8-990b-de4614db2302 req-6755c298-2aae-4d1c-9bc3-7d2a80be733a service nova] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Neutron deleted interface 15a5be02-4db5-4752-ace0-a2e1bcec23c7; detaching it from the instance and deleting it from the info cache [ 1044.925307] env[61906]: DEBUG nova.network.neutron [req-f8f8c64f-6b19-46a8-990b-de4614db2302 req-6755c298-2aae-4d1c-9bc3-7d2a80be733a service nova] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.992043] env[61906]: DEBUG nova.network.neutron [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Successfully updated port: 24ab450e-e9c4-4d9a-9c4d-13dee30eb56b {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1045.059534] env[61906]: DEBUG nova.network.neutron [-] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.304730] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5232b0de-930c-3389-9c81-7ee60567f040, 'name': SearchDatastore_Task, 'duration_secs': 0.011113} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.305514] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5ab9507-0a33-4081-b5b1-1e42d73b090d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.310924] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 1045.310924] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52aed89a-44b7-143c-0cc9-c0ac6523db86" [ 1045.310924] env[61906]: _type = "Task" [ 1045.310924] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.318096] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52aed89a-44b7-143c-0cc9-c0ac6523db86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.414737] env[61906]: DEBUG nova.compute.utils [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1045.418684] env[61906]: DEBUG nova.compute.manager [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1045.419044] env[61906]: DEBUG nova.network.neutron [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1045.427269] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4aa5ba44-dce8-4193-bad1-42a47f8f11ee {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.437552] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fdc1262-833e-4b17-821b-1be109d398c9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.467657] env[61906]: DEBUG nova.compute.manager [req-f8f8c64f-6b19-46a8-990b-de4614db2302 req-6755c298-2aae-4d1c-9bc3-7d2a80be733a service nova] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Detach interface failed, port_id=15a5be02-4db5-4752-ace0-a2e1bcec23c7, reason: Instance 0fe121ad-a539-4c04-bb65-b524cb3d91a8 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1045.467945] env[61906]: DEBUG nova.compute.manager [req-f8f8c64f-6b19-46a8-990b-de4614db2302 req-6755c298-2aae-4d1c-9bc3-7d2a80be733a service nova] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Received event network-vif-plugged-24ab450e-e9c4-4d9a-9c4d-13dee30eb56b {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1045.468216] env[61906]: DEBUG oslo_concurrency.lockutils [req-f8f8c64f-6b19-46a8-990b-de4614db2302 req-6755c298-2aae-4d1c-9bc3-7d2a80be733a service nova] Acquiring lock "86be7197-c0d1-49c5-aa30-cf908a506031-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.468459] env[61906]: DEBUG oslo_concurrency.lockutils [req-f8f8c64f-6b19-46a8-990b-de4614db2302 req-6755c298-2aae-4d1c-9bc3-7d2a80be733a service nova] Lock "86be7197-c0d1-49c5-aa30-cf908a506031-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.468689] env[61906]: DEBUG oslo_concurrency.lockutils [req-f8f8c64f-6b19-46a8-990b-de4614db2302 req-6755c298-2aae-4d1c-9bc3-7d2a80be733a service nova] Lock "86be7197-c0d1-49c5-aa30-cf908a506031-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.468897] env[61906]: DEBUG nova.compute.manager [req-f8f8c64f-6b19-46a8-990b-de4614db2302 req-6755c298-2aae-4d1c-9bc3-7d2a80be733a service nova] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] No waiting events found dispatching network-vif-plugged-24ab450e-e9c4-4d9a-9c4d-13dee30eb56b {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1045.469110] env[61906]: WARNING nova.compute.manager [req-f8f8c64f-6b19-46a8-990b-de4614db2302 req-6755c298-2aae-4d1c-9bc3-7d2a80be733a service nova] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Received unexpected event network-vif-plugged-24ab450e-e9c4-4d9a-9c4d-13dee30eb56b for instance with vm_state building and task_state spawning. [ 1045.481682] env[61906]: DEBUG nova.policy [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '315d6310be014ebdb4b868fd93a680a0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1d6fb2749df6498f82aaf7c5529b309f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1045.493866] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "refresh_cache-86be7197-c0d1-49c5-aa30-cf908a506031" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.493997] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquired lock "refresh_cache-86be7197-c0d1-49c5-aa30-cf908a506031" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.494153] env[61906]: DEBUG nova.network.neutron [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1045.561504] env[61906]: INFO nova.compute.manager [-] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Took 1.21 seconds to deallocate network for instance. [ 1045.592372] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f81c0f2e-15c0-4f02-9c79-51933a5cc702 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.602582] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50abab87-5f83-4b3b-8664-fa9681b8f39e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.633059] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4c30da4-278c-495b-81f4-ff47dd33f28a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.641351] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c05cfd7-1f0b-4379-b64b-2353116fdbf7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.656946] env[61906]: DEBUG nova.compute.provider_tree [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1045.721578] env[61906]: DEBUG nova.network.neutron [req-0cc21e14-0ea5-404b-96ab-0f6a6cafd647 req-a40c5250-4dc9-4074-8f6d-eab55eda93f8 service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Updated VIF entry in instance network info cache for port e0244018-d931-4a96-a25e-228b02a88f56. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1045.721971] env[61906]: DEBUG nova.network.neutron [req-0cc21e14-0ea5-404b-96ab-0f6a6cafd647 req-a40c5250-4dc9-4074-8f6d-eab55eda93f8 service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Updating instance_info_cache with network_info: [{"id": "e0244018-d931-4a96-a25e-228b02a88f56", "address": "fa:16:3e:43:92:23", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0244018-d9", "ovs_interfaceid": "e0244018-d931-4a96-a25e-228b02a88f56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.821853] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52aed89a-44b7-143c-0cc9-c0ac6523db86, 'name': SearchDatastore_Task, 'duration_secs': 0.036986} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.823152] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.823152] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 36c6b1a3-be22-4cec-b21c-0445f5988be1/36c6b1a3-be22-4cec-b21c-0445f5988be1.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1045.823152] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ac4eb7a1-a5de-43d4-a566-c155f35af319 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.829166] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 1045.829166] env[61906]: value = "task-1333635" [ 1045.829166] env[61906]: _type = "Task" [ 1045.829166] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.834290] env[61906]: DEBUG nova.network.neutron [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Successfully created port: 697dbd16-ab06-44ba-ab39-8da5e1ce0c78 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1045.839296] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333635, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.920947] env[61906]: DEBUG nova.compute.manager [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1046.043317] env[61906]: DEBUG nova.network.neutron [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1046.070772] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.160409] env[61906]: DEBUG nova.scheduler.client.report [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1046.215323] env[61906]: DEBUG nova.network.neutron [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Updating instance_info_cache with network_info: [{"id": "24ab450e-e9c4-4d9a-9c4d-13dee30eb56b", "address": "fa:16:3e:fe:51:59", "network": {"id": "9903d8f6-78ef-4ec6-8386-c031ceb715f6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1583036471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d6fb2749df6498f82aaf7c5529b309f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24ab450e-e9", "ovs_interfaceid": "24ab450e-e9c4-4d9a-9c4d-13dee30eb56b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.224417] env[61906]: DEBUG oslo_concurrency.lockutils [req-0cc21e14-0ea5-404b-96ab-0f6a6cafd647 req-a40c5250-4dc9-4074-8f6d-eab55eda93f8 service nova] Releasing lock "refresh_cache-b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.224680] env[61906]: DEBUG nova.compute.manager [req-0cc21e14-0ea5-404b-96ab-0f6a6cafd647 req-a40c5250-4dc9-4074-8f6d-eab55eda93f8 service nova] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Received event network-changed-df3f0533-5c71-429f-9e5b-a39a6eb52ff8 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1046.224858] env[61906]: DEBUG nova.compute.manager [req-0cc21e14-0ea5-404b-96ab-0f6a6cafd647 req-a40c5250-4dc9-4074-8f6d-eab55eda93f8 service nova] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Refreshing instance network info cache due to event network-changed-df3f0533-5c71-429f-9e5b-a39a6eb52ff8. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1046.225081] env[61906]: DEBUG oslo_concurrency.lockutils [req-0cc21e14-0ea5-404b-96ab-0f6a6cafd647 req-a40c5250-4dc9-4074-8f6d-eab55eda93f8 service nova] Acquiring lock "refresh_cache-1be91e16-1acb-4f13-9b26-cae496a0efb8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.225233] env[61906]: DEBUG oslo_concurrency.lockutils [req-0cc21e14-0ea5-404b-96ab-0f6a6cafd647 req-a40c5250-4dc9-4074-8f6d-eab55eda93f8 service nova] Acquired lock "refresh_cache-1be91e16-1acb-4f13-9b26-cae496a0efb8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.225403] env[61906]: DEBUG nova.network.neutron [req-0cc21e14-0ea5-404b-96ab-0f6a6cafd647 req-a40c5250-4dc9-4074-8f6d-eab55eda93f8 service nova] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Refreshing network info cache for port df3f0533-5c71-429f-9e5b-a39a6eb52ff8 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1046.233997] env[61906]: DEBUG nova.network.neutron [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Successfully updated port: a701c10b-b79f-4f4e-a4e0-7d702211ffab {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1046.338961] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333635, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.668635] env[61906]: DEBUG oslo_concurrency.lockutils [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.757s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.671314] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.075s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.671673] env[61906]: DEBUG nova.objects.instance [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lazy-loading 'resources' on Instance uuid ed213a8b-7d7e-47e1-8a4f-d657c13df3ce {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1046.691349] env[61906]: INFO nova.scheduler.client.report [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Deleted allocations for instance b1159533-c970-49d6-af42-b954b20d92fb [ 1046.717439] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Releasing lock "refresh_cache-86be7197-c0d1-49c5-aa30-cf908a506031" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.717762] env[61906]: DEBUG nova.compute.manager [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Instance network_info: |[{"id": "24ab450e-e9c4-4d9a-9c4d-13dee30eb56b", "address": "fa:16:3e:fe:51:59", "network": {"id": "9903d8f6-78ef-4ec6-8386-c031ceb715f6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1583036471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d6fb2749df6498f82aaf7c5529b309f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24ab450e-e9", "ovs_interfaceid": "24ab450e-e9c4-4d9a-9c4d-13dee30eb56b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1046.718201] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fe:51:59', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '184687d6-125a-4b58-bb5b-fdb404088eda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '24ab450e-e9c4-4d9a-9c4d-13dee30eb56b', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1046.725744] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Creating folder: Project (1d6fb2749df6498f82aaf7c5529b309f). Parent ref: group-v284713. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1046.726048] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c384cc20-0ff7-41a4-b634-cdda5ee5c25e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.735989] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "refresh_cache-b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.735989] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired lock "refresh_cache-b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.735989] env[61906]: DEBUG nova.network.neutron [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1046.739532] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Created folder: Project (1d6fb2749df6498f82aaf7c5529b309f) in parent group-v284713. [ 1046.739741] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Creating folder: Instances. Parent ref: group-v284829. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1046.740010] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eb82f518-e203-44a5-8f04-5c5001d748de {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.749475] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Created folder: Instances in parent group-v284829. [ 1046.749786] env[61906]: DEBUG oslo.service.loopingcall [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1046.749993] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1046.750224] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0684c1b4-04ad-40f6-9748-0350837f2db8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.772948] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1046.772948] env[61906]: value = "task-1333638" [ 1046.772948] env[61906]: _type = "Task" [ 1046.772948] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.781076] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333638, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.839996] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333635, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.821923} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.840307] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 36c6b1a3-be22-4cec-b21c-0445f5988be1/36c6b1a3-be22-4cec-b21c-0445f5988be1.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1046.840536] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1046.840829] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-662e81b4-0d9e-4a9b-a45e-0afbde518160 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.848389] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 1046.848389] env[61906]: value = "task-1333639" [ 1046.848389] env[61906]: _type = "Task" [ 1046.848389] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.860314] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333639, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.909153] env[61906]: DEBUG nova.compute.manager [req-ce77be3f-495a-4086-ac94-b48f9acfff33 req-4044a0f1-2798-4759-8c13-c00383dfbd9b service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Received event network-vif-plugged-a701c10b-b79f-4f4e-a4e0-7d702211ffab {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1046.909454] env[61906]: DEBUG oslo_concurrency.lockutils [req-ce77be3f-495a-4086-ac94-b48f9acfff33 req-4044a0f1-2798-4759-8c13-c00383dfbd9b service nova] Acquiring lock "b3e9ab00-8627-4a57-839e-68be5f794b28-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.909741] env[61906]: DEBUG oslo_concurrency.lockutils [req-ce77be3f-495a-4086-ac94-b48f9acfff33 req-4044a0f1-2798-4759-8c13-c00383dfbd9b service nova] Lock "b3e9ab00-8627-4a57-839e-68be5f794b28-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.909994] env[61906]: DEBUG oslo_concurrency.lockutils [req-ce77be3f-495a-4086-ac94-b48f9acfff33 req-4044a0f1-2798-4759-8c13-c00383dfbd9b service nova] Lock "b3e9ab00-8627-4a57-839e-68be5f794b28-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.910251] env[61906]: DEBUG nova.compute.manager [req-ce77be3f-495a-4086-ac94-b48f9acfff33 req-4044a0f1-2798-4759-8c13-c00383dfbd9b service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] No waiting events found dispatching network-vif-plugged-a701c10b-b79f-4f4e-a4e0-7d702211ffab {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1046.910498] env[61906]: WARNING nova.compute.manager [req-ce77be3f-495a-4086-ac94-b48f9acfff33 req-4044a0f1-2798-4759-8c13-c00383dfbd9b service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Received unexpected event network-vif-plugged-a701c10b-b79f-4f4e-a4e0-7d702211ffab for instance with vm_state active and task_state None. [ 1046.910724] env[61906]: DEBUG nova.compute.manager [req-ce77be3f-495a-4086-ac94-b48f9acfff33 req-4044a0f1-2798-4759-8c13-c00383dfbd9b service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Received event network-changed-a701c10b-b79f-4f4e-a4e0-7d702211ffab {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1046.910919] env[61906]: DEBUG nova.compute.manager [req-ce77be3f-495a-4086-ac94-b48f9acfff33 req-4044a0f1-2798-4759-8c13-c00383dfbd9b service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Refreshing instance network info cache due to event network-changed-a701c10b-b79f-4f4e-a4e0-7d702211ffab. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1046.911186] env[61906]: DEBUG oslo_concurrency.lockutils [req-ce77be3f-495a-4086-ac94-b48f9acfff33 req-4044a0f1-2798-4759-8c13-c00383dfbd9b service nova] Acquiring lock "refresh_cache-b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.929789] env[61906]: DEBUG nova.compute.manager [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1046.951032] env[61906]: DEBUG nova.compute.manager [req-27518b5b-267d-454e-8d19-ded89450a7fb req-8859e6b0-ea76-4db3-8536-93f104e19508 service nova] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Received event network-changed-24ab450e-e9c4-4d9a-9c4d-13dee30eb56b {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1046.951268] env[61906]: DEBUG nova.compute.manager [req-27518b5b-267d-454e-8d19-ded89450a7fb req-8859e6b0-ea76-4db3-8536-93f104e19508 service nova] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Refreshing instance network info cache due to event network-changed-24ab450e-e9c4-4d9a-9c4d-13dee30eb56b. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1046.951525] env[61906]: DEBUG oslo_concurrency.lockutils [req-27518b5b-267d-454e-8d19-ded89450a7fb req-8859e6b0-ea76-4db3-8536-93f104e19508 service nova] Acquiring lock "refresh_cache-86be7197-c0d1-49c5-aa30-cf908a506031" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.951729] env[61906]: DEBUG oslo_concurrency.lockutils [req-27518b5b-267d-454e-8d19-ded89450a7fb req-8859e6b0-ea76-4db3-8536-93f104e19508 service nova] Acquired lock "refresh_cache-86be7197-c0d1-49c5-aa30-cf908a506031" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.951931] env[61906]: DEBUG nova.network.neutron [req-27518b5b-267d-454e-8d19-ded89450a7fb req-8859e6b0-ea76-4db3-8536-93f104e19508 service nova] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Refreshing network info cache for port 24ab450e-e9c4-4d9a-9c4d-13dee30eb56b {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1046.956644] env[61906]: DEBUG nova.virt.hardware [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1046.956918] env[61906]: DEBUG nova.virt.hardware [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1046.957113] env[61906]: DEBUG nova.virt.hardware [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1046.957343] env[61906]: DEBUG nova.virt.hardware [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1046.957547] env[61906]: DEBUG nova.virt.hardware [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1046.957753] env[61906]: DEBUG nova.virt.hardware [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1046.958031] env[61906]: DEBUG nova.virt.hardware [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1046.958249] env[61906]: DEBUG nova.virt.hardware [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1046.958475] env[61906]: DEBUG nova.virt.hardware [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1046.958676] env[61906]: DEBUG nova.virt.hardware [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1046.958898] env[61906]: DEBUG nova.virt.hardware [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1046.960011] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed11fe40-d081-4d63-b439-90df24754b3c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.969216] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f41f5e89-6638-4718-871e-b9d47027396f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.029047] env[61906]: DEBUG nova.network.neutron [req-0cc21e14-0ea5-404b-96ab-0f6a6cafd647 req-a40c5250-4dc9-4074-8f6d-eab55eda93f8 service nova] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Updated VIF entry in instance network info cache for port df3f0533-5c71-429f-9e5b-a39a6eb52ff8. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1047.029864] env[61906]: DEBUG nova.network.neutron [req-0cc21e14-0ea5-404b-96ab-0f6a6cafd647 req-a40c5250-4dc9-4074-8f6d-eab55eda93f8 service nova] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Updating instance_info_cache with network_info: [{"id": "df3f0533-5c71-429f-9e5b-a39a6eb52ff8", "address": "fa:16:3e:bc:61:4f", "network": {"id": "c3bcef0f-44a7-44f7-8130-df77c0d3ac2d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-330384233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.134", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "129cdb06b47346b6895df45e73f7b2dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35e463c7-7d78-4d66-8efd-6127b1f3ee17", "external-id": "nsx-vlan-transportzone-175", "segmentation_id": 175, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf3f0533-5c", "ovs_interfaceid": "df3f0533-5c71-429f-9e5b-a39a6eb52ff8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.199667] env[61906]: DEBUG oslo_concurrency.lockutils [None req-060f6b27-a2a1-4c6a-9d86-37c648a38079 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "b1159533-c970-49d6-af42-b954b20d92fb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.650s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.286474] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333638, 'name': CreateVM_Task, 'duration_secs': 0.317296} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.286779] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1047.288098] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.288564] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.289011] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1047.291902] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-733419bd-d859-4d10-84e6-1bf3a2484994 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.298238] env[61906]: WARNING nova.network.neutron [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] 37d8737f-db44-4cf6-8e34-31b4cbd2f82e already exists in list: networks containing: ['37d8737f-db44-4cf6-8e34-31b4cbd2f82e']. ignoring it [ 1047.299690] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1047.299690] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52f15a7b-ccae-3310-362b-f78f1c775b4f" [ 1047.299690] env[61906]: _type = "Task" [ 1047.299690] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.311671] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52f15a7b-ccae-3310-362b-f78f1c775b4f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.315664] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6672b1c5-ec47-4daa-b216-25740a454554 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.323016] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-006fc6d0-f4a7-45fe-8f84-bc7b4dc12e93 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.359554] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e34fd39-52f7-48b1-aa5a-b8e9a48bacb6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.369756] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333639, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062353} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.372625] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1047.373746] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5707aa19-0d5b-421b-957a-6a59229a4f1a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.377565] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8f90cca-46c0-4978-a7c1-b7ae0ef8912e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.403591] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] 36c6b1a3-be22-4cec-b21c-0445f5988be1/36c6b1a3-be22-4cec-b21c-0445f5988be1.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1047.411145] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0a3804c4-5958-4154-af74-69d14a52a189 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.426625] env[61906]: DEBUG nova.compute.provider_tree [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1047.434041] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 1047.434041] env[61906]: value = "task-1333640" [ 1047.434041] env[61906]: _type = "Task" [ 1047.434041] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.442509] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333640, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.484983] env[61906]: DEBUG nova.network.neutron [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Successfully updated port: 697dbd16-ab06-44ba-ab39-8da5e1ce0c78 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1047.535009] env[61906]: DEBUG oslo_concurrency.lockutils [req-0cc21e14-0ea5-404b-96ab-0f6a6cafd647 req-a40c5250-4dc9-4074-8f6d-eab55eda93f8 service nova] Releasing lock "refresh_cache-1be91e16-1acb-4f13-9b26-cae496a0efb8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.678118] env[61906]: DEBUG nova.network.neutron [req-27518b5b-267d-454e-8d19-ded89450a7fb req-8859e6b0-ea76-4db3-8536-93f104e19508 service nova] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Updated VIF entry in instance network info cache for port 24ab450e-e9c4-4d9a-9c4d-13dee30eb56b. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1047.678559] env[61906]: DEBUG nova.network.neutron [req-27518b5b-267d-454e-8d19-ded89450a7fb req-8859e6b0-ea76-4db3-8536-93f104e19508 service nova] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Updating instance_info_cache with network_info: [{"id": "24ab450e-e9c4-4d9a-9c4d-13dee30eb56b", "address": "fa:16:3e:fe:51:59", "network": {"id": "9903d8f6-78ef-4ec6-8386-c031ceb715f6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1583036471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d6fb2749df6498f82aaf7c5529b309f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24ab450e-e9", "ovs_interfaceid": "24ab450e-e9c4-4d9a-9c4d-13dee30eb56b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.770942] env[61906]: DEBUG nova.network.neutron [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Updating instance_info_cache with network_info: [{"id": "e0244018-d931-4a96-a25e-228b02a88f56", "address": "fa:16:3e:43:92:23", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0244018-d9", "ovs_interfaceid": "e0244018-d931-4a96-a25e-228b02a88f56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a701c10b-b79f-4f4e-a4e0-7d702211ffab", "address": "fa:16:3e:d7:47:ab", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa701c10b-b7", "ovs_interfaceid": "a701c10b-b79f-4f4e-a4e0-7d702211ffab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.810445] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52f15a7b-ccae-3310-362b-f78f1c775b4f, 'name': SearchDatastore_Task, 'duration_secs': 0.06404} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.810761] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.810994] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1047.811473] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.811694] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.811930] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1047.812250] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d38e5511-88f6-4132-8b62-29756964f8d0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.826666] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1047.826851] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1047.827599] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b705b3cb-b64c-4e96-8764-106f1c381495 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.832687] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1047.832687] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]522e8c90-3f44-d7c7-90d8-668a6fd6906f" [ 1047.832687] env[61906]: _type = "Task" [ 1047.832687] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.841252] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]522e8c90-3f44-d7c7-90d8-668a6fd6906f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.930414] env[61906]: DEBUG nova.scheduler.client.report [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1047.943769] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333640, 'name': ReconfigVM_Task, 'duration_secs': 0.296256} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.944584] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Reconfigured VM instance instance-00000063 to attach disk [datastore2] 36c6b1a3-be22-4cec-b21c-0445f5988be1/36c6b1a3-be22-4cec-b21c-0445f5988be1.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1047.945226] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-03b7faaa-8926-4176-8f64-ea849cecfdd4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.951676] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 1047.951676] env[61906]: value = "task-1333641" [ 1047.951676] env[61906]: _type = "Task" [ 1047.951676] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.959290] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333641, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.987997] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "refresh_cache-f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.988183] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquired lock "refresh_cache-f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.988392] env[61906]: DEBUG nova.network.neutron [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1048.183822] env[61906]: DEBUG oslo_concurrency.lockutils [req-27518b5b-267d-454e-8d19-ded89450a7fb req-8859e6b0-ea76-4db3-8536-93f104e19508 service nova] Releasing lock "refresh_cache-86be7197-c0d1-49c5-aa30-cf908a506031" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.270553] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Releasing lock "refresh_cache-b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.270938] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.271136] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired lock "b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.271522] env[61906]: DEBUG oslo_concurrency.lockutils [req-ce77be3f-495a-4086-ac94-b48f9acfff33 req-4044a0f1-2798-4759-8c13-c00383dfbd9b service nova] Acquired lock "refresh_cache-b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.271644] env[61906]: DEBUG nova.network.neutron [req-ce77be3f-495a-4086-ac94-b48f9acfff33 req-4044a0f1-2798-4759-8c13-c00383dfbd9b service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Refreshing network info cache for port a701c10b-b79f-4f4e-a4e0-7d702211ffab {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1048.273391] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad8f8d3e-d125-4e2e-b72a-902f613cbc70 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.290438] env[61906]: DEBUG nova.virt.hardware [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1048.290665] env[61906]: DEBUG nova.virt.hardware [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1048.290830] env[61906]: DEBUG nova.virt.hardware [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1048.291028] env[61906]: DEBUG nova.virt.hardware [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1048.291185] env[61906]: DEBUG nova.virt.hardware [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1048.291373] env[61906]: DEBUG nova.virt.hardware [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1048.291645] env[61906]: DEBUG nova.virt.hardware [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1048.291812] env[61906]: DEBUG nova.virt.hardware [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1048.291935] env[61906]: DEBUG nova.virt.hardware [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1048.292123] env[61906]: DEBUG nova.virt.hardware [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1048.292309] env[61906]: DEBUG nova.virt.hardware [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1048.298534] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Reconfiguring VM to attach interface {{(pid=61906) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1048.299453] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-08cfa35f-9262-4eb2-8f02-7f5856f4d2d0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.317531] env[61906]: DEBUG oslo_vmware.api [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 1048.317531] env[61906]: value = "task-1333642" [ 1048.317531] env[61906]: _type = "Task" [ 1048.317531] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.325822] env[61906]: DEBUG oslo_vmware.api [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333642, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.341972] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]522e8c90-3f44-d7c7-90d8-668a6fd6906f, 'name': SearchDatastore_Task, 'duration_secs': 0.054867} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.342750] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a14159ea-cb61-4af4-8b4e-490a9bea73bc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.347563] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1048.347563] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]521332ed-cdfa-01c8-490c-81b324a775aa" [ 1048.347563] env[61906]: _type = "Task" [ 1048.347563] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.355302] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]521332ed-cdfa-01c8-490c-81b324a775aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.435930] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.764s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.438584] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.368s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.438805] env[61906]: DEBUG nova.objects.instance [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lazy-loading 'resources' on Instance uuid 0fe121ad-a539-4c04-bb65-b524cb3d91a8 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1048.462208] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333641, 'name': Rename_Task, 'duration_secs': 0.135912} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.463289] env[61906]: INFO nova.scheduler.client.report [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Deleted allocations for instance ed213a8b-7d7e-47e1-8a4f-d657c13df3ce [ 1048.464305] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1048.466936] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8666bb1b-812a-4d80-b715-6a6995ba9189 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.473160] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 1048.473160] env[61906]: value = "task-1333643" [ 1048.473160] env[61906]: _type = "Task" [ 1048.473160] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.481181] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333643, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.523143] env[61906]: DEBUG nova.network.neutron [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1048.719341] env[61906]: DEBUG nova.network.neutron [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Updating instance_info_cache with network_info: [{"id": "697dbd16-ab06-44ba-ab39-8da5e1ce0c78", "address": "fa:16:3e:80:02:ba", "network": {"id": "9903d8f6-78ef-4ec6-8386-c031ceb715f6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1583036471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d6fb2749df6498f82aaf7c5529b309f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap697dbd16-ab", "ovs_interfaceid": "697dbd16-ab06-44ba-ab39-8da5e1ce0c78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.832907] env[61906]: DEBUG oslo_vmware.api [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333642, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.860300] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]521332ed-cdfa-01c8-490c-81b324a775aa, 'name': SearchDatastore_Task, 'duration_secs': 0.011531} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.860609] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.860831] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 86be7197-c0d1-49c5-aa30-cf908a506031/86be7197-c0d1-49c5-aa30-cf908a506031.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1048.861119] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6127b402-5505-4970-a0dd-4278b680dc7a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.867733] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1048.867733] env[61906]: value = "task-1333644" [ 1048.867733] env[61906]: _type = "Task" [ 1048.867733] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.878115] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333644, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.974532] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0c31f7ed-9237-43da-90f7-0de9e2be48c3 tempest-ServerDiskConfigTestJSON-789598290 tempest-ServerDiskConfigTestJSON-789598290-project-member] Lock "ed213a8b-7d7e-47e1-8a4f-d657c13df3ce" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.293s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.982141] env[61906]: DEBUG nova.compute.manager [req-82d458ad-7a32-42bf-b90a-21395ea626e3 req-d3b4b090-369f-4394-bf70-8b20be8068f4 service nova] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Received event network-vif-plugged-697dbd16-ab06-44ba-ab39-8da5e1ce0c78 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1048.982302] env[61906]: DEBUG oslo_concurrency.lockutils [req-82d458ad-7a32-42bf-b90a-21395ea626e3 req-d3b4b090-369f-4394-bf70-8b20be8068f4 service nova] Acquiring lock "f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.982572] env[61906]: DEBUG oslo_concurrency.lockutils [req-82d458ad-7a32-42bf-b90a-21395ea626e3 req-d3b4b090-369f-4394-bf70-8b20be8068f4 service nova] Lock "f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.982776] env[61906]: DEBUG oslo_concurrency.lockutils [req-82d458ad-7a32-42bf-b90a-21395ea626e3 req-d3b4b090-369f-4394-bf70-8b20be8068f4 service nova] Lock "f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.983017] env[61906]: DEBUG nova.compute.manager [req-82d458ad-7a32-42bf-b90a-21395ea626e3 req-d3b4b090-369f-4394-bf70-8b20be8068f4 service nova] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] No waiting events found dispatching network-vif-plugged-697dbd16-ab06-44ba-ab39-8da5e1ce0c78 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1048.983207] env[61906]: WARNING nova.compute.manager [req-82d458ad-7a32-42bf-b90a-21395ea626e3 req-d3b4b090-369f-4394-bf70-8b20be8068f4 service nova] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Received unexpected event network-vif-plugged-697dbd16-ab06-44ba-ab39-8da5e1ce0c78 for instance with vm_state building and task_state spawning. [ 1048.983836] env[61906]: DEBUG nova.compute.manager [req-82d458ad-7a32-42bf-b90a-21395ea626e3 req-d3b4b090-369f-4394-bf70-8b20be8068f4 service nova] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Received event network-changed-697dbd16-ab06-44ba-ab39-8da5e1ce0c78 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1048.983836] env[61906]: DEBUG nova.compute.manager [req-82d458ad-7a32-42bf-b90a-21395ea626e3 req-d3b4b090-369f-4394-bf70-8b20be8068f4 service nova] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Refreshing instance network info cache due to event network-changed-697dbd16-ab06-44ba-ab39-8da5e1ce0c78. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1048.983836] env[61906]: DEBUG oslo_concurrency.lockutils [req-82d458ad-7a32-42bf-b90a-21395ea626e3 req-d3b4b090-369f-4394-bf70-8b20be8068f4 service nova] Acquiring lock "refresh_cache-f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.987750] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333643, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.001517] env[61906]: DEBUG nova.network.neutron [req-ce77be3f-495a-4086-ac94-b48f9acfff33 req-4044a0f1-2798-4759-8c13-c00383dfbd9b service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Updated VIF entry in instance network info cache for port a701c10b-b79f-4f4e-a4e0-7d702211ffab. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1049.002039] env[61906]: DEBUG nova.network.neutron [req-ce77be3f-495a-4086-ac94-b48f9acfff33 req-4044a0f1-2798-4759-8c13-c00383dfbd9b service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Updating instance_info_cache with network_info: [{"id": "e0244018-d931-4a96-a25e-228b02a88f56", "address": "fa:16:3e:43:92:23", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0244018-d9", "ovs_interfaceid": "e0244018-d931-4a96-a25e-228b02a88f56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a701c10b-b79f-4f4e-a4e0-7d702211ffab", "address": "fa:16:3e:d7:47:ab", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa701c10b-b7", "ovs_interfaceid": "a701c10b-b79f-4f4e-a4e0-7d702211ffab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.077782] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97bda67d-0668-43b1-819b-46aa66513cb8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.085365] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1020d8ab-ff0e-42c1-bfce-9b381a1e15d5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.116396] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a55671-b407-45ba-8fea-19cbf61559f6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.122016] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "256457e7-8d94-46f8-8edb-c5782ac34225" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.122273] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "256457e7-8d94-46f8-8edb-c5782ac34225" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.128571] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31a292c9-8a9d-4170-88a6-1daad6e12350 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.145747] env[61906]: DEBUG nova.compute.provider_tree [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1049.221945] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Releasing lock "refresh_cache-f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.222351] env[61906]: DEBUG nova.compute.manager [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Instance network_info: |[{"id": "697dbd16-ab06-44ba-ab39-8da5e1ce0c78", "address": "fa:16:3e:80:02:ba", "network": {"id": "9903d8f6-78ef-4ec6-8386-c031ceb715f6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1583036471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d6fb2749df6498f82aaf7c5529b309f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap697dbd16-ab", "ovs_interfaceid": "697dbd16-ab06-44ba-ab39-8da5e1ce0c78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1049.222684] env[61906]: DEBUG oslo_concurrency.lockutils [req-82d458ad-7a32-42bf-b90a-21395ea626e3 req-d3b4b090-369f-4394-bf70-8b20be8068f4 service nova] Acquired lock "refresh_cache-f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.222874] env[61906]: DEBUG nova.network.neutron [req-82d458ad-7a32-42bf-b90a-21395ea626e3 req-d3b4b090-369f-4394-bf70-8b20be8068f4 service nova] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Refreshing network info cache for port 697dbd16-ab06-44ba-ab39-8da5e1ce0c78 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1049.224354] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:80:02:ba', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '184687d6-125a-4b58-bb5b-fdb404088eda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '697dbd16-ab06-44ba-ab39-8da5e1ce0c78', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1049.232033] env[61906]: DEBUG oslo.service.loopingcall [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1049.232970] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1049.233565] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-12afb567-9212-4d6b-aa88-b74f245f1985 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.253931] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1049.253931] env[61906]: value = "task-1333645" [ 1049.253931] env[61906]: _type = "Task" [ 1049.253931] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.262874] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333645, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.330975] env[61906]: DEBUG oslo_vmware.api [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333642, 'name': ReconfigVM_Task, 'duration_secs': 0.574917} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.331593] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Releasing lock "b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.332097] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Reconfigured VM to attach interface {{(pid=61906) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1049.378158] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333644, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.484191] env[61906]: DEBUG oslo_vmware.api [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333643, 'name': PowerOnVM_Task, 'duration_secs': 0.66115} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.484561] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1049.484666] env[61906]: INFO nova.compute.manager [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Took 9.32 seconds to spawn the instance on the hypervisor. [ 1049.484860] env[61906]: DEBUG nova.compute.manager [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1049.485667] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-634507d1-a662-427a-b851-d2afe09be837 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.507777] env[61906]: DEBUG oslo_concurrency.lockutils [req-ce77be3f-495a-4086-ac94-b48f9acfff33 req-4044a0f1-2798-4759-8c13-c00383dfbd9b service nova] Releasing lock "refresh_cache-b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.624659] env[61906]: DEBUG nova.compute.manager [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1049.649365] env[61906]: DEBUG nova.scheduler.client.report [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1049.765664] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333645, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.836764] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2a3a584a-d9c9-45ee-aee8-c5529bf660ed tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "interface-b3e9ab00-8627-4a57-839e-68be5f794b28-a701c10b-b79f-4f4e-a4e0-7d702211ffab" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.435s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.881190] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333644, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.627402} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.881504] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 86be7197-c0d1-49c5-aa30-cf908a506031/86be7197-c0d1-49c5-aa30-cf908a506031.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1049.881741] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1049.882007] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-68ce6100-6f70-4f9b-b935-8b7a2db6fa99 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.888901] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1049.888901] env[61906]: value = "task-1333646" [ 1049.888901] env[61906]: _type = "Task" [ 1049.888901] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.898546] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333646, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.005075] env[61906]: INFO nova.compute.manager [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Took 14.85 seconds to build instance. [ 1050.148209] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.154466] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.716s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.156610] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.009s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.161733] env[61906]: INFO nova.compute.claims [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1050.176876] env[61906]: DEBUG nova.network.neutron [req-82d458ad-7a32-42bf-b90a-21395ea626e3 req-d3b4b090-369f-4394-bf70-8b20be8068f4 service nova] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Updated VIF entry in instance network info cache for port 697dbd16-ab06-44ba-ab39-8da5e1ce0c78. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1050.179587] env[61906]: DEBUG nova.network.neutron [req-82d458ad-7a32-42bf-b90a-21395ea626e3 req-d3b4b090-369f-4394-bf70-8b20be8068f4 service nova] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Updating instance_info_cache with network_info: [{"id": "697dbd16-ab06-44ba-ab39-8da5e1ce0c78", "address": "fa:16:3e:80:02:ba", "network": {"id": "9903d8f6-78ef-4ec6-8386-c031ceb715f6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1583036471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d6fb2749df6498f82aaf7c5529b309f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap697dbd16-ab", "ovs_interfaceid": "697dbd16-ab06-44ba-ab39-8da5e1ce0c78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.181618] env[61906]: INFO nova.scheduler.client.report [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Deleted allocations for instance 0fe121ad-a539-4c04-bb65-b524cb3d91a8 [ 1050.264253] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333645, 'name': CreateVM_Task, 'duration_secs': 0.557093} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.264424] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1050.265106] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.265292] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.265638] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1050.266179] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6dfb9ae-f134-4d2b-aa86-be2357c5d04b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.270202] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1050.270202] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52b5d815-a118-f353-df1b-4b5169c3d88d" [ 1050.270202] env[61906]: _type = "Task" [ 1050.270202] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.277564] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52b5d815-a118-f353-df1b-4b5169c3d88d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.400769] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333646, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070958} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.406134] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1050.407009] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f6fb1ed-8b2a-4684-99b9-7f7140831328 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.434033] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 86be7197-c0d1-49c5-aa30-cf908a506031/86be7197-c0d1-49c5-aa30-cf908a506031.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1050.434217] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f1deeca6-066a-40fc-9d7f-ed168b064e16 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.456117] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1050.456117] env[61906]: value = "task-1333647" [ 1050.456117] env[61906]: _type = "Task" [ 1050.456117] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.463994] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333647, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.507440] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b99a667-62d4-495d-8bce-cea3dca8ed39 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "36c6b1a3-be22-4cec-b21c-0445f5988be1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.361s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.684577] env[61906]: DEBUG oslo_concurrency.lockutils [req-82d458ad-7a32-42bf-b90a-21395ea626e3 req-d3b4b090-369f-4394-bf70-8b20be8068f4 service nova] Releasing lock "refresh_cache-f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.689626] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9eb21da9-6f7f-4be8-ad29-ae24cf40258d tempest-ImagesTestJSON-1332288202 tempest-ImagesTestJSON-1332288202-project-member] Lock "0fe121ad-a539-4c04-bb65-b524cb3d91a8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.507s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.782455] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52b5d815-a118-f353-df1b-4b5169c3d88d, 'name': SearchDatastore_Task, 'duration_secs': 0.009237} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.783032] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.783442] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1050.785212] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.785212] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.785212] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1050.785212] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ad8dbbca-7399-4132-a8ec-289101657179 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.795205] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1050.795205] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1050.795205] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c16db7d-6c17-489a-98be-7bd5e12198b1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.801769] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1050.801769] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52953ab0-6a74-9444-0649-cb8a423eee83" [ 1050.801769] env[61906]: _type = "Task" [ 1050.801769] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.813453] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52953ab0-6a74-9444-0649-cb8a423eee83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.973805] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333647, 'name': ReconfigVM_Task, 'duration_secs': 0.333149} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.973805] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 86be7197-c0d1-49c5-aa30-cf908a506031/86be7197-c0d1-49c5-aa30-cf908a506031.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1050.973805] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9e53147e-12dc-4fb0-87bd-c7e483ebd1f0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.984287] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1050.984287] env[61906]: value = "task-1333648" [ 1050.984287] env[61906]: _type = "Task" [ 1050.984287] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.993890] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333648, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.308278] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a03bc37a-8725-4de8-a483-2024a361967f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.319062] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d57daa98-55b3-4482-a0ac-1564d5502a45 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.324078] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52953ab0-6a74-9444-0649-cb8a423eee83, 'name': SearchDatastore_Task, 'duration_secs': 0.009933} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.324078] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "interface-b3e9ab00-8627-4a57-839e-68be5f794b28-a701c10b-b79f-4f4e-a4e0-7d702211ffab" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.324267] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "interface-b3e9ab00-8627-4a57-839e-68be5f794b28-a701c10b-b79f-4f4e-a4e0-7d702211ffab" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.325532] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0ca5580-0f71-4705-8219-7f126858d67c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.356307] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01894533-a6f2-4ede-a62b-7ce78a1fac3b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.360495] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1051.360495] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52eea8e8-7f97-ee47-dcb3-0d5a2e72fb0d" [ 1051.360495] env[61906]: _type = "Task" [ 1051.360495] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.367410] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce11eb48-1992-4350-aa7b-c61c5bc0ef73 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.375018] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52eea8e8-7f97-ee47-dcb3-0d5a2e72fb0d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.384756] env[61906]: DEBUG nova.compute.provider_tree [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1051.495722] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333648, 'name': Rename_Task, 'duration_secs': 0.145576} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.495722] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1051.495722] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c78fbf46-e3ed-40ad-91c5-363fc330845d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.502253] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1051.502253] env[61906]: value = "task-1333649" [ 1051.502253] env[61906]: _type = "Task" [ 1051.502253] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.509986] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333649, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.566886] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc3c21be-2170-410d-9346-f8157e81656c tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "36c6b1a3-be22-4cec-b21c-0445f5988be1" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.567217] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc3c21be-2170-410d-9346-f8157e81656c tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "36c6b1a3-be22-4cec-b21c-0445f5988be1" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.567435] env[61906]: DEBUG nova.compute.manager [None req-dc3c21be-2170-410d-9346-f8157e81656c tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1051.568408] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc15c093-996f-48b6-b423-6b3fc55aae0d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.575896] env[61906]: DEBUG nova.compute.manager [None req-dc3c21be-2170-410d-9346-f8157e81656c tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61906) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1051.576825] env[61906]: DEBUG nova.objects.instance [None req-dc3c21be-2170-410d-9346-f8157e81656c tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lazy-loading 'flavor' on Instance uuid 36c6b1a3-be22-4cec-b21c-0445f5988be1 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1051.830401] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.830640] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired lock "b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.831653] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-911b0061-685e-413c-be25-0ca6e08b9614 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.851461] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "216bd7f2-dd02-4d3d-a00c-15f62b3de7ca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.851745] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "216bd7f2-dd02-4d3d-a00c-15f62b3de7ca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.857018] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db4c6a11-be36-4e27-adfc-85bff008936d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.881480] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Reconfiguring VM to detach interface {{(pid=61906) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1051.885647] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-808ca4b6-cc69-4408-adde-c75bf0137a7e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.900173] env[61906]: DEBUG nova.scheduler.client.report [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1051.909315] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52eea8e8-7f97-ee47-dcb3-0d5a2e72fb0d, 'name': SearchDatastore_Task, 'duration_secs': 0.013541} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.911457] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.911723] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b/f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1051.911955] env[61906]: DEBUG oslo_vmware.api [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 1051.911955] env[61906]: value = "task-1333650" [ 1051.911955] env[61906]: _type = "Task" [ 1051.911955] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.912188] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e51dcec4-c45d-4492-beb7-0111a91b4c9d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.925539] env[61906]: DEBUG oslo_vmware.api [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333650, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.926566] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1051.926566] env[61906]: value = "task-1333651" [ 1051.926566] env[61906]: _type = "Task" [ 1051.926566] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.934987] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333651, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.011795] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333649, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.083008] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc3c21be-2170-410d-9346-f8157e81656c tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1052.083328] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e0aede84-a7e0-4bd7-83af-77169f074d8b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.091292] env[61906]: DEBUG oslo_vmware.api [None req-dc3c21be-2170-410d-9346-f8157e81656c tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 1052.091292] env[61906]: value = "task-1333652" [ 1052.091292] env[61906]: _type = "Task" [ 1052.091292] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.100931] env[61906]: DEBUG oslo_vmware.api [None req-dc3c21be-2170-410d-9346-f8157e81656c tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333652, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.357499] env[61906]: DEBUG nova.compute.manager [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1052.406270] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.249s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.407112] env[61906]: DEBUG nova.compute.manager [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1052.436027] env[61906]: DEBUG oslo_vmware.api [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333650, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.442030] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333651, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.512777] env[61906]: DEBUG oslo_vmware.api [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333649, 'name': PowerOnVM_Task, 'duration_secs': 0.63347} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.513106] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1052.513297] env[61906]: INFO nova.compute.manager [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Took 7.88 seconds to spawn the instance on the hypervisor. [ 1052.513509] env[61906]: DEBUG nova.compute.manager [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1052.514287] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67fb28bc-6abe-48a9-8308-a5b3d78bb5bb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.601167] env[61906]: DEBUG oslo_vmware.api [None req-dc3c21be-2170-410d-9346-f8157e81656c tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333652, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.896134] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.896134] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.897394] env[61906]: INFO nova.compute.claims [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1052.912480] env[61906]: DEBUG nova.compute.utils [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1052.913992] env[61906]: DEBUG nova.compute.manager [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1052.914188] env[61906]: DEBUG nova.network.neutron [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1052.929422] env[61906]: DEBUG oslo_vmware.api [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333650, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.938684] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333651, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.573595} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.940357] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b/f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1052.940357] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1052.940357] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9c0079aa-1521-4963-babf-6bf583c542e8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.947708] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1052.947708] env[61906]: value = "task-1333653" [ 1052.947708] env[61906]: _type = "Task" [ 1052.947708] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.965624] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333653, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.010204] env[61906]: DEBUG nova.policy [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fd25149493b948f2826e6688c9958e6a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4fbdcd2dc0d04e83b9f82a19bdfd0023', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1053.030260] env[61906]: INFO nova.compute.manager [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Took 12.68 seconds to build instance. [ 1053.105861] env[61906]: DEBUG oslo_vmware.api [None req-dc3c21be-2170-410d-9346-f8157e81656c tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333652, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.417575] env[61906]: DEBUG nova.compute.manager [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1053.431591] env[61906]: DEBUG oslo_vmware.api [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333650, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.458790] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333653, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.198738} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.462066] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1053.462066] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58911654-ba82-41aa-bd1b-60cdc465c8d2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.485445] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b/f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1053.486709] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd675642-c865-4eb7-b89d-78e668afa5b3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.511271] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1053.511271] env[61906]: value = "task-1333654" [ 1053.511271] env[61906]: _type = "Task" [ 1053.511271] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.520564] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333654, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.539553] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aee71d74-9b10-4974-839b-4d253d3dd3f7 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "86be7197-c0d1-49c5-aa30-cf908a506031" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.196s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.585891] env[61906]: DEBUG nova.network.neutron [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Successfully created port: 261ecf70-132a-4fe2-be6b-556a555c86cb {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1053.605355] env[61906]: DEBUG oslo_vmware.api [None req-dc3c21be-2170-410d-9346-f8157e81656c tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333652, 'name': PowerOffVM_Task, 'duration_secs': 1.324875} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.605664] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc3c21be-2170-410d-9346-f8157e81656c tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1053.606113] env[61906]: DEBUG nova.compute.manager [None req-dc3c21be-2170-410d-9346-f8157e81656c tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1053.608958] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81e1f42d-07df-4f71-b01c-8ee543998491 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.936282] env[61906]: DEBUG oslo_vmware.api [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333650, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.024161] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333654, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.078978] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf1a69bf-ac65-42a3-80e7-d2a21b30099b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.087261] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fc8770b-8287-47a5-9096-bb8524910a17 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.121943] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95220df8-8753-47e8-8785-cb46db7080cf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.125466] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dc3c21be-2170-410d-9346-f8157e81656c tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "36c6b1a3-be22-4cec-b21c-0445f5988be1" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.558s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.131870] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd5ec233-5ca6-4302-827d-38dcf9a1143c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.147800] env[61906]: DEBUG nova.compute.provider_tree [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1054.434677] env[61906]: DEBUG nova.compute.manager [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1054.444190] env[61906]: DEBUG oslo_vmware.api [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333650, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.464020] env[61906]: DEBUG nova.virt.hardware [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1054.464020] env[61906]: DEBUG nova.virt.hardware [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1054.464020] env[61906]: DEBUG nova.virt.hardware [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1054.464020] env[61906]: DEBUG nova.virt.hardware [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1054.464020] env[61906]: DEBUG nova.virt.hardware [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1054.464020] env[61906]: DEBUG nova.virt.hardware [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1054.464354] env[61906]: DEBUG nova.virt.hardware [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1054.464651] env[61906]: DEBUG nova.virt.hardware [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1054.464976] env[61906]: DEBUG nova.virt.hardware [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1054.465300] env[61906]: DEBUG nova.virt.hardware [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1054.465598] env[61906]: DEBUG nova.virt.hardware [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1054.466579] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15989f7c-8f18-4b75-9ca9-f6480b571555 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.475486] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-963e2609-6114-4355-835a-cc6dc908adbf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.522352] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333654, 'name': ReconfigVM_Task, 'duration_secs': 0.651394} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.525478] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Reconfigured VM instance instance-00000065 to attach disk [datastore2] f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b/f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1054.525478] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b80c7195-182f-447f-9d01-82fd09a4806c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.533018] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1054.533018] env[61906]: value = "task-1333655" [ 1054.533018] env[61906]: _type = "Task" [ 1054.533018] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.538744] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333655, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.651530] env[61906]: DEBUG nova.scheduler.client.report [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1054.938313] env[61906]: DEBUG oslo_vmware.api [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333650, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.041782] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333655, 'name': Rename_Task, 'duration_secs': 0.140078} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.042619] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1055.042926] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5cbffa5a-7964-4ad6-8256-0b4c2b4348c4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.051946] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1055.051946] env[61906]: value = "task-1333656" [ 1055.051946] env[61906]: _type = "Task" [ 1055.051946] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.060922] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333656, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.095853] env[61906]: DEBUG nova.compute.manager [req-e23b6824-52b3-4353-9864-7684d354cece req-8eb6d663-7db3-472e-9c48-735745924811 service nova] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Received event network-vif-plugged-261ecf70-132a-4fe2-be6b-556a555c86cb {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1055.096021] env[61906]: DEBUG oslo_concurrency.lockutils [req-e23b6824-52b3-4353-9864-7684d354cece req-8eb6d663-7db3-472e-9c48-735745924811 service nova] Acquiring lock "256457e7-8d94-46f8-8edb-c5782ac34225-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.096208] env[61906]: DEBUG oslo_concurrency.lockutils [req-e23b6824-52b3-4353-9864-7684d354cece req-8eb6d663-7db3-472e-9c48-735745924811 service nova] Lock "256457e7-8d94-46f8-8edb-c5782ac34225-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.096363] env[61906]: DEBUG oslo_concurrency.lockutils [req-e23b6824-52b3-4353-9864-7684d354cece req-8eb6d663-7db3-472e-9c48-735745924811 service nova] Lock "256457e7-8d94-46f8-8edb-c5782ac34225-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.096532] env[61906]: DEBUG nova.compute.manager [req-e23b6824-52b3-4353-9864-7684d354cece req-8eb6d663-7db3-472e-9c48-735745924811 service nova] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] No waiting events found dispatching network-vif-plugged-261ecf70-132a-4fe2-be6b-556a555c86cb {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1055.096706] env[61906]: WARNING nova.compute.manager [req-e23b6824-52b3-4353-9864-7684d354cece req-8eb6d663-7db3-472e-9c48-735745924811 service nova] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Received unexpected event network-vif-plugged-261ecf70-132a-4fe2-be6b-556a555c86cb for instance with vm_state building and task_state spawning. [ 1055.158167] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.263s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.158630] env[61906]: DEBUG nova.compute.manager [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1055.217177] env[61906]: DEBUG nova.network.neutron [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Successfully updated port: 261ecf70-132a-4fe2-be6b-556a555c86cb {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1055.439175] env[61906]: DEBUG oslo_vmware.api [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333650, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.564706] env[61906]: DEBUG oslo_vmware.api [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333656, 'name': PowerOnVM_Task, 'duration_secs': 0.450163} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.564993] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1055.565333] env[61906]: INFO nova.compute.manager [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Took 8.64 seconds to spawn the instance on the hypervisor. [ 1055.565590] env[61906]: DEBUG nova.compute.manager [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1055.567697] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a7f0702-222b-4d1f-9cb9-a29aa07f047d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.663534] env[61906]: DEBUG nova.compute.utils [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1055.665614] env[61906]: DEBUG nova.compute.manager [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1055.665807] env[61906]: DEBUG nova.network.neutron [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1055.709574] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "36c6b1a3-be22-4cec-b21c-0445f5988be1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.710047] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "36c6b1a3-be22-4cec-b21c-0445f5988be1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.710392] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "36c6b1a3-be22-4cec-b21c-0445f5988be1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.710705] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "36c6b1a3-be22-4cec-b21c-0445f5988be1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.710935] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "36c6b1a3-be22-4cec-b21c-0445f5988be1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.714014] env[61906]: DEBUG nova.policy [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '27ba0d3b00ce4513a03259970a600190', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b7d3aa88849f4492aa16ff337b8cfdda', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1055.715905] env[61906]: INFO nova.compute.manager [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Terminating instance [ 1055.718566] env[61906]: DEBUG nova.compute.manager [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1055.718739] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1055.719233] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "refresh_cache-256457e7-8d94-46f8-8edb-c5782ac34225" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.719360] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquired lock "refresh_cache-256457e7-8d94-46f8-8edb-c5782ac34225" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.719518] env[61906]: DEBUG nova.network.neutron [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1055.720942] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d2a230e-a6c0-4112-b0be-b16755f9e004 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.730210] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1055.730491] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f2f2a52c-3cc7-451d-aaeb-99c936ac5405 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.939222] env[61906]: DEBUG oslo_vmware.api [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333650, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.984043] env[61906]: DEBUG nova.network.neutron [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Successfully created port: 8716273f-1918-42d4-9166-9788da93cecd {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1056.039328] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1056.039665] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1056.039935] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Deleting the datastore file [datastore2] 36c6b1a3-be22-4cec-b21c-0445f5988be1 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1056.040315] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-449df2f2-176e-4db6-bc62-784afa545300 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.048695] env[61906]: DEBUG oslo_vmware.api [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for the task: (returnval){ [ 1056.048695] env[61906]: value = "task-1333658" [ 1056.048695] env[61906]: _type = "Task" [ 1056.048695] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.059946] env[61906]: DEBUG oslo_vmware.api [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333658, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.083741] env[61906]: INFO nova.compute.manager [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Took 15.23 seconds to build instance. [ 1056.170013] env[61906]: DEBUG nova.compute.manager [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1056.254909] env[61906]: DEBUG nova.network.neutron [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1056.402133] env[61906]: DEBUG nova.network.neutron [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Updating instance_info_cache with network_info: [{"id": "261ecf70-132a-4fe2-be6b-556a555c86cb", "address": "fa:16:3e:84:13:a4", "network": {"id": "1c55d0cc-4c30-4b89-8736-3e96c5c9c507", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-245940583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fbdcd2dc0d04e83b9f82a19bdfd0023", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bcd9d2d-25c8-41ad-9a4a-93b9029ba993", "external-id": "nsx-vlan-transportzone-734", "segmentation_id": 734, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap261ecf70-13", "ovs_interfaceid": "261ecf70-132a-4fe2-be6b-556a555c86cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.440343] env[61906]: DEBUG oslo_vmware.api [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333650, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.559688] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.560133] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.564921] env[61906]: DEBUG oslo_vmware.api [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Task: {'id': task-1333658, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.362505} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.565609] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1056.565609] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1056.565772] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1056.566143] env[61906]: INFO nova.compute.manager [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Took 0.85 seconds to destroy the instance on the hypervisor. [ 1056.566143] env[61906]: DEBUG oslo.service.loopingcall [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1056.566391] env[61906]: DEBUG nova.compute.manager [-] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1056.566481] env[61906]: DEBUG nova.network.neutron [-] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1056.586499] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ab82d19d-025c-41a5-8131-6a554ebb942d tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.745s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.869209] env[61906]: DEBUG nova.compute.manager [req-8884fc0f-2f58-4e44-af18-ab4a4ed18460 req-697ad9ea-0e05-416a-bd84-68b3eea3c53b service nova] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Received event network-vif-deleted-e2c74014-2d68-4807-86d9-3009eee2db02 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1056.869209] env[61906]: INFO nova.compute.manager [req-8884fc0f-2f58-4e44-af18-ab4a4ed18460 req-697ad9ea-0e05-416a-bd84-68b3eea3c53b service nova] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Neutron deleted interface e2c74014-2d68-4807-86d9-3009eee2db02; detaching it from the instance and deleting it from the info cache [ 1056.869209] env[61906]: DEBUG nova.network.neutron [req-8884fc0f-2f58-4e44-af18-ab4a4ed18460 req-697ad9ea-0e05-416a-bd84-68b3eea3c53b service nova] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.904327] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Releasing lock "refresh_cache-256457e7-8d94-46f8-8edb-c5782ac34225" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.904673] env[61906]: DEBUG nova.compute.manager [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Instance network_info: |[{"id": "261ecf70-132a-4fe2-be6b-556a555c86cb", "address": "fa:16:3e:84:13:a4", "network": {"id": "1c55d0cc-4c30-4b89-8736-3e96c5c9c507", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-245940583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fbdcd2dc0d04e83b9f82a19bdfd0023", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bcd9d2d-25c8-41ad-9a4a-93b9029ba993", "external-id": "nsx-vlan-transportzone-734", "segmentation_id": 734, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap261ecf70-13", "ovs_interfaceid": "261ecf70-132a-4fe2-be6b-556a555c86cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1056.905104] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:84:13:a4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7bcd9d2d-25c8-41ad-9a4a-93b9029ba993', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '261ecf70-132a-4fe2-be6b-556a555c86cb', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1056.912784] env[61906]: DEBUG oslo.service.loopingcall [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1056.913227] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1056.913497] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-38417f13-8ea7-40f4-94ef-1f4a4061d30c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.935248] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1056.935248] env[61906]: value = "task-1333659" [ 1056.935248] env[61906]: _type = "Task" [ 1056.935248] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.941136] env[61906]: DEBUG oslo_vmware.api [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333650, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.945920] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333659, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.067046] env[61906]: DEBUG nova.compute.manager [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1057.121830] env[61906]: DEBUG nova.compute.manager [req-b47100e5-5b8c-49b6-96c2-2e9d20976e12 req-5a121a5d-8f21-4e70-8981-cb834bb845b6 service nova] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Received event network-changed-261ecf70-132a-4fe2-be6b-556a555c86cb {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1057.121997] env[61906]: DEBUG nova.compute.manager [req-b47100e5-5b8c-49b6-96c2-2e9d20976e12 req-5a121a5d-8f21-4e70-8981-cb834bb845b6 service nova] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Refreshing instance network info cache due to event network-changed-261ecf70-132a-4fe2-be6b-556a555c86cb. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1057.122238] env[61906]: DEBUG oslo_concurrency.lockutils [req-b47100e5-5b8c-49b6-96c2-2e9d20976e12 req-5a121a5d-8f21-4e70-8981-cb834bb845b6 service nova] Acquiring lock "refresh_cache-256457e7-8d94-46f8-8edb-c5782ac34225" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.122323] env[61906]: DEBUG oslo_concurrency.lockutils [req-b47100e5-5b8c-49b6-96c2-2e9d20976e12 req-5a121a5d-8f21-4e70-8981-cb834bb845b6 service nova] Acquired lock "refresh_cache-256457e7-8d94-46f8-8edb-c5782ac34225" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.122487] env[61906]: DEBUG nova.network.neutron [req-b47100e5-5b8c-49b6-96c2-2e9d20976e12 req-5a121a5d-8f21-4e70-8981-cb834bb845b6 service nova] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Refreshing network info cache for port 261ecf70-132a-4fe2-be6b-556a555c86cb {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1057.137648] env[61906]: INFO nova.compute.manager [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Rescuing [ 1057.137923] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "refresh_cache-f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.138095] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquired lock "refresh_cache-f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.138271] env[61906]: DEBUG nova.network.neutron [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1057.185177] env[61906]: DEBUG nova.compute.manager [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1057.210751] env[61906]: DEBUG nova.virt.hardware [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1057.211036] env[61906]: DEBUG nova.virt.hardware [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1057.211218] env[61906]: DEBUG nova.virt.hardware [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1057.211443] env[61906]: DEBUG nova.virt.hardware [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1057.212933] env[61906]: DEBUG nova.virt.hardware [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1057.212933] env[61906]: DEBUG nova.virt.hardware [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1057.212933] env[61906]: DEBUG nova.virt.hardware [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1057.212933] env[61906]: DEBUG nova.virt.hardware [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1057.212933] env[61906]: DEBUG nova.virt.hardware [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1057.212933] env[61906]: DEBUG nova.virt.hardware [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1057.212933] env[61906]: DEBUG nova.virt.hardware [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1057.213809] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae9b5611-225f-405f-9be9-754fbb69d7b0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.221894] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e4736ab-a258-4b28-8b62-4ab07cdec108 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.349294] env[61906]: DEBUG nova.network.neutron [-] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.371577] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ff2c0d26-38af-4621-b517-08c1ae1d42bc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.381053] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d502224-f186-4b01-9dc0-4cd2ff56fca2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.407395] env[61906]: DEBUG nova.compute.manager [req-8884fc0f-2f58-4e44-af18-ab4a4ed18460 req-697ad9ea-0e05-416a-bd84-68b3eea3c53b service nova] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Detach interface failed, port_id=e2c74014-2d68-4807-86d9-3009eee2db02, reason: Instance 36c6b1a3-be22-4cec-b21c-0445f5988be1 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1057.441979] env[61906]: DEBUG oslo_vmware.api [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333650, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.446811] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333659, 'name': CreateVM_Task, 'duration_secs': 0.344257} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.446977] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1057.447660] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.447819] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.448162] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1057.448421] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75c49e06-e207-43ea-aee5-45a570055627 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.452633] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 1057.452633] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52991324-14a4-88b3-53db-34a3cb731973" [ 1057.452633] env[61906]: _type = "Task" [ 1057.452633] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.456067] env[61906]: DEBUG nova.network.neutron [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Successfully updated port: 8716273f-1918-42d4-9166-9788da93cecd {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1057.461729] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52991324-14a4-88b3-53db-34a3cb731973, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.592602] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.592909] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.594868] env[61906]: INFO nova.compute.claims [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1057.855090] env[61906]: INFO nova.compute.manager [-] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Took 1.29 seconds to deallocate network for instance. [ 1057.885549] env[61906]: DEBUG nova.network.neutron [req-b47100e5-5b8c-49b6-96c2-2e9d20976e12 req-5a121a5d-8f21-4e70-8981-cb834bb845b6 service nova] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Updated VIF entry in instance network info cache for port 261ecf70-132a-4fe2-be6b-556a555c86cb. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1057.889017] env[61906]: DEBUG nova.network.neutron [req-b47100e5-5b8c-49b6-96c2-2e9d20976e12 req-5a121a5d-8f21-4e70-8981-cb834bb845b6 service nova] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Updating instance_info_cache with network_info: [{"id": "261ecf70-132a-4fe2-be6b-556a555c86cb", "address": "fa:16:3e:84:13:a4", "network": {"id": "1c55d0cc-4c30-4b89-8736-3e96c5c9c507", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-245940583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fbdcd2dc0d04e83b9f82a19bdfd0023", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bcd9d2d-25c8-41ad-9a4a-93b9029ba993", "external-id": "nsx-vlan-transportzone-734", "segmentation_id": 734, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap261ecf70-13", "ovs_interfaceid": "261ecf70-132a-4fe2-be6b-556a555c86cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.898983] env[61906]: DEBUG nova.network.neutron [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Updating instance_info_cache with network_info: [{"id": "697dbd16-ab06-44ba-ab39-8da5e1ce0c78", "address": "fa:16:3e:80:02:ba", "network": {"id": "9903d8f6-78ef-4ec6-8386-c031ceb715f6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1583036471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d6fb2749df6498f82aaf7c5529b309f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap697dbd16-ab", "ovs_interfaceid": "697dbd16-ab06-44ba-ab39-8da5e1ce0c78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.941379] env[61906]: DEBUG oslo_vmware.api [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333650, 'name': ReconfigVM_Task, 'duration_secs': 5.758643} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.942139] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Releasing lock "b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.942139] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Reconfigured VM to detach interface {{(pid=61906) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1057.963600] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "refresh_cache-216bd7f2-dd02-4d3d-a00c-15f62b3de7ca" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.963785] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired lock "refresh_cache-216bd7f2-dd02-4d3d-a00c-15f62b3de7ca" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.963933] env[61906]: DEBUG nova.network.neutron [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1057.965042] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52991324-14a4-88b3-53db-34a3cb731973, 'name': SearchDatastore_Task, 'duration_secs': 0.011336} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.965337] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.965558] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1057.966670] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.966670] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.966670] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1057.966670] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-37dcf7ad-150d-49b9-9bca-d4555b4d78f9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.975256] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1057.975471] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1057.976614] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76d4da33-95ea-4459-9a8d-7b7ecdb3eca2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.982439] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 1057.982439] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52eee081-624b-1172-69f4-32f24fe283dd" [ 1057.982439] env[61906]: _type = "Task" [ 1057.982439] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.990761] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52eee081-624b-1172-69f4-32f24fe283dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.361114] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.388696] env[61906]: DEBUG oslo_concurrency.lockutils [req-b47100e5-5b8c-49b6-96c2-2e9d20976e12 req-5a121a5d-8f21-4e70-8981-cb834bb845b6 service nova] Releasing lock "refresh_cache-256457e7-8d94-46f8-8edb-c5782ac34225" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.401704] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Releasing lock "refresh_cache-f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.493551] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52eee081-624b-1172-69f4-32f24fe283dd, 'name': SearchDatastore_Task, 'duration_secs': 0.010123} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.494679] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc9e6c7f-31a1-4cd2-bfca-609976d0cc1d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.497824] env[61906]: DEBUG nova.network.neutron [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1058.502939] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 1058.502939] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5212ebe0-0b1c-6183-18ab-d4defa15bbcf" [ 1058.502939] env[61906]: _type = "Task" [ 1058.502939] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.510419] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5212ebe0-0b1c-6183-18ab-d4defa15bbcf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.632800] env[61906]: DEBUG nova.network.neutron [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Updating instance_info_cache with network_info: [{"id": "8716273f-1918-42d4-9166-9788da93cecd", "address": "fa:16:3e:e8:71:41", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8716273f-19", "ovs_interfaceid": "8716273f-1918-42d4-9166-9788da93cecd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.734496] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-663de3a0-e05f-4a47-995b-8ce489fd3238 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.742275] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1257995-2545-4d3b-b209-accd034b0962 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.773942] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49140bac-047d-45d5-9adb-19d46e80e6f3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.782314] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7753b35-e2cd-42dd-81db-9b32adeaf2fe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.797813] env[61906]: DEBUG nova.compute.provider_tree [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1058.934189] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1058.934529] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-225c6a6c-fdc2-4507-b156-76dc9288f83c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.948699] env[61906]: DEBUG oslo_vmware.api [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1058.948699] env[61906]: value = "task-1333660" [ 1058.948699] env[61906]: _type = "Task" [ 1058.948699] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.965387] env[61906]: DEBUG oslo_vmware.api [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333660, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.012973] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5212ebe0-0b1c-6183-18ab-d4defa15bbcf, 'name': SearchDatastore_Task, 'duration_secs': 0.035217} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.013277] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.013594] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 256457e7-8d94-46f8-8edb-c5782ac34225/256457e7-8d94-46f8-8edb-c5782ac34225.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1059.013897] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-849800fc-7d2b-4d33-9a75-c2082c580917 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.020848] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 1059.020848] env[61906]: value = "task-1333661" [ 1059.020848] env[61906]: _type = "Task" [ 1059.020848] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.029120] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333661, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.138451] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Releasing lock "refresh_cache-216bd7f2-dd02-4d3d-a00c-15f62b3de7ca" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.139205] env[61906]: DEBUG nova.compute.manager [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Instance network_info: |[{"id": "8716273f-1918-42d4-9166-9788da93cecd", "address": "fa:16:3e:e8:71:41", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8716273f-19", "ovs_interfaceid": "8716273f-1918-42d4-9166-9788da93cecd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1059.139377] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e8:71:41', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8716273f-1918-42d4-9166-9788da93cecd', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1059.147533] env[61906]: DEBUG oslo.service.loopingcall [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1059.149339] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1059.150215] env[61906]: DEBUG nova.compute.manager [req-2b677aee-206c-4e44-b11d-3c7d8586d691 req-aefb1a0a-682b-42a9-8e53-f6fc2d4363fe service nova] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Received event network-vif-plugged-8716273f-1918-42d4-9166-9788da93cecd {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1059.150413] env[61906]: DEBUG oslo_concurrency.lockutils [req-2b677aee-206c-4e44-b11d-3c7d8586d691 req-aefb1a0a-682b-42a9-8e53-f6fc2d4363fe service nova] Acquiring lock "216bd7f2-dd02-4d3d-a00c-15f62b3de7ca-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.150628] env[61906]: DEBUG oslo_concurrency.lockutils [req-2b677aee-206c-4e44-b11d-3c7d8586d691 req-aefb1a0a-682b-42a9-8e53-f6fc2d4363fe service nova] Lock "216bd7f2-dd02-4d3d-a00c-15f62b3de7ca-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.150798] env[61906]: DEBUG oslo_concurrency.lockutils [req-2b677aee-206c-4e44-b11d-3c7d8586d691 req-aefb1a0a-682b-42a9-8e53-f6fc2d4363fe service nova] Lock "216bd7f2-dd02-4d3d-a00c-15f62b3de7ca-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.150967] env[61906]: DEBUG nova.compute.manager [req-2b677aee-206c-4e44-b11d-3c7d8586d691 req-aefb1a0a-682b-42a9-8e53-f6fc2d4363fe service nova] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] No waiting events found dispatching network-vif-plugged-8716273f-1918-42d4-9166-9788da93cecd {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1059.151156] env[61906]: WARNING nova.compute.manager [req-2b677aee-206c-4e44-b11d-3c7d8586d691 req-aefb1a0a-682b-42a9-8e53-f6fc2d4363fe service nova] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Received unexpected event network-vif-plugged-8716273f-1918-42d4-9166-9788da93cecd for instance with vm_state building and task_state spawning. [ 1059.151318] env[61906]: DEBUG nova.compute.manager [req-2b677aee-206c-4e44-b11d-3c7d8586d691 req-aefb1a0a-682b-42a9-8e53-f6fc2d4363fe service nova] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Received event network-changed-8716273f-1918-42d4-9166-9788da93cecd {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1059.151548] env[61906]: DEBUG nova.compute.manager [req-2b677aee-206c-4e44-b11d-3c7d8586d691 req-aefb1a0a-682b-42a9-8e53-f6fc2d4363fe service nova] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Refreshing instance network info cache due to event network-changed-8716273f-1918-42d4-9166-9788da93cecd. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1059.151769] env[61906]: DEBUG oslo_concurrency.lockutils [req-2b677aee-206c-4e44-b11d-3c7d8586d691 req-aefb1a0a-682b-42a9-8e53-f6fc2d4363fe service nova] Acquiring lock "refresh_cache-216bd7f2-dd02-4d3d-a00c-15f62b3de7ca" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.151914] env[61906]: DEBUG oslo_concurrency.lockutils [req-2b677aee-206c-4e44-b11d-3c7d8586d691 req-aefb1a0a-682b-42a9-8e53-f6fc2d4363fe service nova] Acquired lock "refresh_cache-216bd7f2-dd02-4d3d-a00c-15f62b3de7ca" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.152097] env[61906]: DEBUG nova.network.neutron [req-2b677aee-206c-4e44-b11d-3c7d8586d691 req-aefb1a0a-682b-42a9-8e53-f6fc2d4363fe service nova] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Refreshing network info cache for port 8716273f-1918-42d4-9166-9788da93cecd {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1059.153645] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7383bbaf-2be0-4d7b-8ac6-6baaa04fd0dd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.176014] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1059.176014] env[61906]: value = "task-1333662" [ 1059.176014] env[61906]: _type = "Task" [ 1059.176014] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.181964] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333662, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.273679] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "refresh_cache-b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.273968] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquired lock "refresh_cache-b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.274259] env[61906]: DEBUG nova.network.neutron [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1059.301086] env[61906]: DEBUG nova.scheduler.client.report [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1059.459804] env[61906]: DEBUG oslo_vmware.api [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333660, 'name': PowerOffVM_Task, 'duration_secs': 0.224684} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.460052] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1059.460889] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30f5574f-1c3d-4ddd-a4fd-e3799c58d0d2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.480881] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d146795-886c-4624-a11b-a3dba10d2b68 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.511778] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1059.512195] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7c78ec5d-3a69-4011-bdca-f7ae59d1b643 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.520657] env[61906]: DEBUG oslo_vmware.api [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1059.520657] env[61906]: value = "task-1333663" [ 1059.520657] env[61906]: _type = "Task" [ 1059.520657] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.537289] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] VM already powered off {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1059.537624] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1059.537929] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.538111] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.538333] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1059.538656] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333661, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.538908] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bfa3589a-a5ee-4002-9443-103d9e76a241 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.549448] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1059.549699] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1059.550565] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14e22f41-79d2-47ea-b292-f5137679f642 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.556720] env[61906]: DEBUG oslo_vmware.api [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1059.556720] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52e4a8e5-c446-5281-7b7f-c050d8be55d6" [ 1059.556720] env[61906]: _type = "Task" [ 1059.556720] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.565356] env[61906]: DEBUG oslo_vmware.api [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52e4a8e5-c446-5281-7b7f-c050d8be55d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.684336] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333662, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.800708] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "b3e9ab00-8627-4a57-839e-68be5f794b28" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.800959] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "b3e9ab00-8627-4a57-839e-68be5f794b28" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.801185] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "b3e9ab00-8627-4a57-839e-68be5f794b28-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.801398] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "b3e9ab00-8627-4a57-839e-68be5f794b28-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.801540] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "b3e9ab00-8627-4a57-839e-68be5f794b28-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.804875] env[61906]: INFO nova.compute.manager [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Terminating instance [ 1059.806614] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.214s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.807098] env[61906]: DEBUG nova.compute.manager [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1059.809790] env[61906]: DEBUG nova.compute.manager [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1059.809989] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1059.812254] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.451s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.812473] env[61906]: DEBUG nova.objects.instance [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lazy-loading 'resources' on Instance uuid 36c6b1a3-be22-4cec-b21c-0445f5988be1 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1059.813827] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8cece77-121b-4f4d-969e-61b6eafdb624 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.822776] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1059.823567] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aaeffa1e-dc48-41b7-a0cf-1f89fb1d626b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.831689] env[61906]: DEBUG oslo_vmware.api [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 1059.831689] env[61906]: value = "task-1333664" [ 1059.831689] env[61906]: _type = "Task" [ 1059.831689] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.840991] env[61906]: DEBUG oslo_vmware.api [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333664, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.922258] env[61906]: DEBUG nova.network.neutron [req-2b677aee-206c-4e44-b11d-3c7d8586d691 req-aefb1a0a-682b-42a9-8e53-f6fc2d4363fe service nova] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Updated VIF entry in instance network info cache for port 8716273f-1918-42d4-9166-9788da93cecd. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1059.922664] env[61906]: DEBUG nova.network.neutron [req-2b677aee-206c-4e44-b11d-3c7d8586d691 req-aefb1a0a-682b-42a9-8e53-f6fc2d4363fe service nova] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Updating instance_info_cache with network_info: [{"id": "8716273f-1918-42d4-9166-9788da93cecd", "address": "fa:16:3e:e8:71:41", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8716273f-19", "ovs_interfaceid": "8716273f-1918-42d4-9166-9788da93cecd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.024442] env[61906]: INFO nova.network.neutron [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Port a701c10b-b79f-4f4e-a4e0-7d702211ffab from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1060.024884] env[61906]: DEBUG nova.network.neutron [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Updating instance_info_cache with network_info: [{"id": "e0244018-d931-4a96-a25e-228b02a88f56", "address": "fa:16:3e:43:92:23", "network": {"id": "37d8737f-db44-4cf6-8e34-31b4cbd2f82e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-370626072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36ff7a500ef444cbf5a168c5a48208b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35ac9709-fd8b-4630-897a-68ed629d1b11", "external-id": "nsx-vlan-transportzone-284", "segmentation_id": 284, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0244018-d9", "ovs_interfaceid": "e0244018-d931-4a96-a25e-228b02a88f56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.035379] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333661, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.598998} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.036168] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 256457e7-8d94-46f8-8edb-c5782ac34225/256457e7-8d94-46f8-8edb-c5782ac34225.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1060.036396] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1060.036651] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d0e6dba2-716e-4144-b440-c3f7540c9f3d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.043594] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 1060.043594] env[61906]: value = "task-1333665" [ 1060.043594] env[61906]: _type = "Task" [ 1060.043594] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.051897] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333665, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.066264] env[61906]: DEBUG oslo_vmware.api [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52e4a8e5-c446-5281-7b7f-c050d8be55d6, 'name': SearchDatastore_Task, 'duration_secs': 0.016732} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.067061] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aec0ea69-9ac4-4842-9b28-4eb9fc281821 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.072607] env[61906]: DEBUG oslo_vmware.api [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1060.072607] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52ce230a-dfd6-d53f-143e-358577a62c85" [ 1060.072607] env[61906]: _type = "Task" [ 1060.072607] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.080181] env[61906]: DEBUG oslo_vmware.api [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52ce230a-dfd6-d53f-143e-358577a62c85, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.186779] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333662, 'name': CreateVM_Task, 'duration_secs': 0.609833} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.186959] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1060.187646] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.187824] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.188420] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1060.188420] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec191491-876e-46a7-8a85-66c1e17531d5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.192817] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1060.192817] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5249e16b-6887-c6e8-b593-1c792b0cb37c" [ 1060.192817] env[61906]: _type = "Task" [ 1060.192817] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.200290] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5249e16b-6887-c6e8-b593-1c792b0cb37c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.318526] env[61906]: DEBUG nova.compute.utils [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1060.322642] env[61906]: DEBUG nova.compute.manager [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1060.322818] env[61906]: DEBUG nova.network.neutron [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1060.341207] env[61906]: DEBUG oslo_vmware.api [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333664, 'name': PowerOffVM_Task, 'duration_secs': 0.193366} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.343716] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1060.343899] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1060.344345] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-29d1031a-9533-460d-9377-496cf9258749 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.363044] env[61906]: DEBUG nova.policy [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3896807bf4d64d08b988b6c2023a408b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '00c03ca0de1447da9d47c83f256efdcd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1060.408942] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1060.409193] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1060.409381] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Deleting the datastore file [datastore1] b3e9ab00-8627-4a57-839e-68be5f794b28 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1060.411788] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8c7bc57d-30b8-498c-be64-1a995ce713cb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.418146] env[61906]: DEBUG oslo_vmware.api [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 1060.418146] env[61906]: value = "task-1333667" [ 1060.418146] env[61906]: _type = "Task" [ 1060.418146] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.428830] env[61906]: DEBUG oslo_concurrency.lockutils [req-2b677aee-206c-4e44-b11d-3c7d8586d691 req-aefb1a0a-682b-42a9-8e53-f6fc2d4363fe service nova] Releasing lock "refresh_cache-216bd7f2-dd02-4d3d-a00c-15f62b3de7ca" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.429523] env[61906]: DEBUG oslo_vmware.api [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333667, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.460472] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cff6c3bb-c8b2-4943-949a-11ab9dd50b28 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.468332] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-868e9a16-82e9-40d2-8003-0188751dd867 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.499987] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a24d72-4f69-4aed-8fdc-66fdff311145 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.508188] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9af1bc-6f7d-46b9-bd33-16a59938553c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.522077] env[61906]: DEBUG nova.compute.provider_tree [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1060.530487] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Releasing lock "refresh_cache-b3e9ab00-8627-4a57-839e-68be5f794b28" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.562706] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333665, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063378} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.562884] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1060.564159] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2c12550-ddc7-4fe4-82ba-083989f08395 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.590922] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 256457e7-8d94-46f8-8edb-c5782ac34225/256457e7-8d94-46f8-8edb-c5782ac34225.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1060.594727] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d0553e93-ab15-4899-a903-4ea8b5cf466a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.614719] env[61906]: DEBUG oslo_vmware.api [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52ce230a-dfd6-d53f-143e-358577a62c85, 'name': SearchDatastore_Task, 'duration_secs': 0.00855} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.617352] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.617352] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b/a9502563-1f96-46a5-b87e-12c0e9375e6e-rescue.vmdk. {{(pid=61906) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1060.617352] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 1060.617352] env[61906]: value = "task-1333668" [ 1060.617352] env[61906]: _type = "Task" [ 1060.617352] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.617352] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-86418549-73ea-4e52-9ebe-52010a4d2a0e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.628155] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333668, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.629563] env[61906]: DEBUG oslo_vmware.api [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1060.629563] env[61906]: value = "task-1333669" [ 1060.629563] env[61906]: _type = "Task" [ 1060.629563] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.634852] env[61906]: DEBUG nova.network.neutron [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Successfully created port: 1d0e39a6-9bf4-4034-b7e3-0d500971fc09 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1060.639699] env[61906]: DEBUG oslo_vmware.api [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333669, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.703383] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5249e16b-6887-c6e8-b593-1c792b0cb37c, 'name': SearchDatastore_Task, 'duration_secs': 0.008784} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.703791] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.704092] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1060.704409] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.704608] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.704844] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1060.705160] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-49e208f7-53e6-4ba6-9c00-db17d8318392 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.715932] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1060.716187] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1060.716969] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8db53b65-b000-42dd-aa87-14e1bfe523c4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.722407] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1060.722407] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52e730d6-b32e-0688-a8a9-520534adccae" [ 1060.722407] env[61906]: _type = "Task" [ 1060.722407] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.730320] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52e730d6-b32e-0688-a8a9-520534adccae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.823797] env[61906]: DEBUG nova.compute.manager [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1060.928426] env[61906]: DEBUG oslo_vmware.api [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333667, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157594} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.928779] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1060.929074] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1060.929464] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1060.929607] env[61906]: INFO nova.compute.manager [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1060.929929] env[61906]: DEBUG oslo.service.loopingcall [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1060.930165] env[61906]: DEBUG nova.compute.manager [-] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1060.930310] env[61906]: DEBUG nova.network.neutron [-] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1061.025143] env[61906]: DEBUG nova.scheduler.client.report [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1061.035575] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f9f07458-9885-4ed9-b1e6-231650f71412 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "interface-b3e9ab00-8627-4a57-839e-68be5f794b28-a701c10b-b79f-4f4e-a4e0-7d702211ffab" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.711s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.128641] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333668, 'name': ReconfigVM_Task, 'duration_secs': 0.354526} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.128945] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 256457e7-8d94-46f8-8edb-c5782ac34225/256457e7-8d94-46f8-8edb-c5782ac34225.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1061.129597] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2ece49fc-e28a-4092-a8d4-26ba963f4ac9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.139321] env[61906]: DEBUG oslo_vmware.api [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333669, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.140530] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 1061.140530] env[61906]: value = "task-1333670" [ 1061.140530] env[61906]: _type = "Task" [ 1061.140530] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.151761] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333670, 'name': Rename_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.234583] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52e730d6-b32e-0688-a8a9-520534adccae, 'name': SearchDatastore_Task, 'duration_secs': 0.00919} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.236013] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8d4a8f1-2f66-4414-b02e-96bbe1bd9b83 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.241682] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1061.241682] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5200a2d5-bd3a-0725-9730-9bcd27c56ac7" [ 1061.241682] env[61906]: _type = "Task" [ 1061.241682] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.248262] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5200a2d5-bd3a-0725-9730-9bcd27c56ac7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.532101] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.718s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.553743] env[61906]: INFO nova.scheduler.client.report [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Deleted allocations for instance 36c6b1a3-be22-4cec-b21c-0445f5988be1 [ 1061.642989] env[61906]: DEBUG oslo_vmware.api [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333669, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.566169} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.647074] env[61906]: INFO nova.virt.vmwareapi.ds_util [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b/a9502563-1f96-46a5-b87e-12c0e9375e6e-rescue.vmdk. [ 1061.648314] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20aa2fd7-5c5a-498a-8c7f-879ac278a334 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.655791] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333670, 'name': Rename_Task, 'duration_secs': 0.487211} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.669469] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1061.676869] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b/a9502563-1f96-46a5-b87e-12c0e9375e6e-rescue.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1061.677271] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bc90c0cb-68e7-4592-a6d9-d453d213a314 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.678709] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be96aea7-9aaa-47ef-b91d-5f3e3686ded2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.696594] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 1061.696594] env[61906]: value = "task-1333671" [ 1061.696594] env[61906]: _type = "Task" [ 1061.696594] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.697800] env[61906]: DEBUG oslo_vmware.api [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1061.697800] env[61906]: value = "task-1333672" [ 1061.697800] env[61906]: _type = "Task" [ 1061.697800] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.707858] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333671, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.710947] env[61906]: DEBUG oslo_vmware.api [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333672, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.752097] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5200a2d5-bd3a-0725-9730-9bcd27c56ac7, 'name': SearchDatastore_Task, 'duration_secs': 0.079208} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.752335] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.752648] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca/216bd7f2-dd02-4d3d-a00c-15f62b3de7ca.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1061.752954] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-210fcfad-56f7-46d0-8611-93a056a9f374 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.760699] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1061.760699] env[61906]: value = "task-1333673" [ 1061.760699] env[61906]: _type = "Task" [ 1061.760699] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.770125] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333673, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.787164] env[61906]: DEBUG nova.compute.manager [req-9d4e4250-5a66-4a68-ab0b-d4e088c11ec5 req-6a0a6c7d-8e79-469b-a3ed-bd64a7d46667 service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Received event network-vif-deleted-e0244018-d931-4a96-a25e-228b02a88f56 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1061.787387] env[61906]: INFO nova.compute.manager [req-9d4e4250-5a66-4a68-ab0b-d4e088c11ec5 req-6a0a6c7d-8e79-469b-a3ed-bd64a7d46667 service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Neutron deleted interface e0244018-d931-4a96-a25e-228b02a88f56; detaching it from the instance and deleting it from the info cache [ 1061.787816] env[61906]: DEBUG nova.network.neutron [req-9d4e4250-5a66-4a68-ab0b-d4e088c11ec5 req-6a0a6c7d-8e79-469b-a3ed-bd64a7d46667 service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.833443] env[61906]: DEBUG nova.compute.manager [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1061.858684] env[61906]: DEBUG nova.virt.hardware [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1061.858990] env[61906]: DEBUG nova.virt.hardware [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1061.859152] env[61906]: DEBUG nova.virt.hardware [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1061.859351] env[61906]: DEBUG nova.virt.hardware [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1061.859504] env[61906]: DEBUG nova.virt.hardware [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1061.859700] env[61906]: DEBUG nova.virt.hardware [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1061.859931] env[61906]: DEBUG nova.virt.hardware [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1061.860163] env[61906]: DEBUG nova.virt.hardware [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1061.860335] env[61906]: DEBUG nova.virt.hardware [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1061.860512] env[61906]: DEBUG nova.virt.hardware [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1061.860697] env[61906]: DEBUG nova.virt.hardware [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1061.861614] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-245658a8-76f5-458b-b43f-4f240a231e13 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.870156] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f36c2fa8-ab8b-44c3-bca4-62ca6846a2ed {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.057163] env[61906]: DEBUG nova.compute.manager [req-177cc774-c76a-4980-a0b1-87073dd2ce05 req-52f2461c-950b-4b39-8260-c9eea8dd4e8c service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Received event network-vif-plugged-1d0e39a6-9bf4-4034-b7e3-0d500971fc09 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1062.057163] env[61906]: DEBUG oslo_concurrency.lockutils [req-177cc774-c76a-4980-a0b1-87073dd2ce05 req-52f2461c-950b-4b39-8260-c9eea8dd4e8c service nova] Acquiring lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.057163] env[61906]: DEBUG oslo_concurrency.lockutils [req-177cc774-c76a-4980-a0b1-87073dd2ce05 req-52f2461c-950b-4b39-8260-c9eea8dd4e8c service nova] Lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.057163] env[61906]: DEBUG oslo_concurrency.lockutils [req-177cc774-c76a-4980-a0b1-87073dd2ce05 req-52f2461c-950b-4b39-8260-c9eea8dd4e8c service nova] Lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.057163] env[61906]: DEBUG nova.compute.manager [req-177cc774-c76a-4980-a0b1-87073dd2ce05 req-52f2461c-950b-4b39-8260-c9eea8dd4e8c service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] No waiting events found dispatching network-vif-plugged-1d0e39a6-9bf4-4034-b7e3-0d500971fc09 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1062.058229] env[61906]: WARNING nova.compute.manager [req-177cc774-c76a-4980-a0b1-87073dd2ce05 req-52f2461c-950b-4b39-8260-c9eea8dd4e8c service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Received unexpected event network-vif-plugged-1d0e39a6-9bf4-4034-b7e3-0d500971fc09 for instance with vm_state building and task_state spawning. [ 1062.062845] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ce2b680-8caf-4044-951d-96ac5d9e2ab8 tempest-ServersTestJSON-2046954064 tempest-ServersTestJSON-2046954064-project-member] Lock "36c6b1a3-be22-4cec-b21c-0445f5988be1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.353s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.203927] env[61906]: DEBUG nova.network.neutron [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Successfully updated port: 1d0e39a6-9bf4-4034-b7e3-0d500971fc09 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1062.218033] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333671, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.220796] env[61906]: DEBUG oslo_vmware.api [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333672, 'name': ReconfigVM_Task, 'duration_secs': 0.320488} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.221476] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Reconfigured VM instance instance-00000065 to attach disk [datastore2] f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b/a9502563-1f96-46a5-b87e-12c0e9375e6e-rescue.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1062.222723] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fa8f34e-70be-462b-bfb2-e8a93600080f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.251427] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e1425879-dcb4-4158-9372-8405480b2b03 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.267200] env[61906]: DEBUG nova.network.neutron [-] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.274731] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333673, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.276477] env[61906]: DEBUG oslo_vmware.api [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1062.276477] env[61906]: value = "task-1333674" [ 1062.276477] env[61906]: _type = "Task" [ 1062.276477] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.286814] env[61906]: DEBUG oslo_vmware.api [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333674, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.289930] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-69cb5506-c2c0-43a1-b78c-50291fe04766 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.303150] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f2ee6e9-47d8-4f67-819b-92ad33b48be3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.334085] env[61906]: DEBUG nova.compute.manager [req-9d4e4250-5a66-4a68-ab0b-d4e088c11ec5 req-6a0a6c7d-8e79-469b-a3ed-bd64a7d46667 service nova] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Detach interface failed, port_id=e0244018-d931-4a96-a25e-228b02a88f56, reason: Instance b3e9ab00-8627-4a57-839e-68be5f794b28 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1062.708547] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "refresh_cache-151cc6f4-96c5-4ac3-97d5-73cddfa7b465" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.708810] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquired lock "refresh_cache-151cc6f4-96c5-4ac3-97d5-73cddfa7b465" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.708841] env[61906]: DEBUG nova.network.neutron [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1062.710333] env[61906]: DEBUG oslo_vmware.api [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333671, 'name': PowerOnVM_Task, 'duration_secs': 0.938485} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.711037] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1062.711274] env[61906]: INFO nova.compute.manager [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Took 8.28 seconds to spawn the instance on the hypervisor. [ 1062.711519] env[61906]: DEBUG nova.compute.manager [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1062.712250] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12c5c283-93db-4869-ae9e-577ce38b8928 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.774290] env[61906]: INFO nova.compute.manager [-] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Took 1.84 seconds to deallocate network for instance. [ 1062.775598] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333673, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.656105} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.776462] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca/216bd7f2-dd02-4d3d-a00c-15f62b3de7ca.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1062.776692] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1062.776987] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d725161b-8073-46e9-8c21-204824dae6ec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.790194] env[61906]: DEBUG oslo_vmware.api [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333674, 'name': ReconfigVM_Task, 'duration_secs': 0.237004} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.791611] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1062.791941] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1062.791941] env[61906]: value = "task-1333675" [ 1062.791941] env[61906]: _type = "Task" [ 1062.791941] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.792148] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-df2fe1bb-ddab-4749-a6eb-0ede228503b3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.804274] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333675, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.805621] env[61906]: DEBUG oslo_vmware.api [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1062.805621] env[61906]: value = "task-1333676" [ 1062.805621] env[61906]: _type = "Task" [ 1062.805621] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.813999] env[61906]: DEBUG oslo_vmware.api [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333676, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.229292] env[61906]: INFO nova.compute.manager [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Took 13.10 seconds to build instance. [ 1063.242201] env[61906]: DEBUG nova.network.neutron [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1063.289446] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.289720] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.289953] env[61906]: DEBUG nova.objects.instance [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lazy-loading 'resources' on Instance uuid b3e9ab00-8627-4a57-839e-68be5f794b28 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1063.305861] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333675, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071107} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.308417] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1063.312086] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03d4f5d5-78d7-4431-86cb-d30e8d955e42 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.319457] env[61906]: DEBUG oslo_vmware.api [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333676, 'name': PowerOnVM_Task, 'duration_secs': 0.423497} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.328926] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1063.339477] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca/216bd7f2-dd02-4d3d-a00c-15f62b3de7ca.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1063.340568] env[61906]: DEBUG nova.compute.manager [None req-8d02e859-e526-4072-80c7-6b6b06e02545 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1063.340827] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34930ad4-febb-4ae0-af01-fd54b7c2281e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.355507] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bc3497d-aefd-43fc-ba0a-c86b27f30b20 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.366398] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1063.366398] env[61906]: value = "task-1333677" [ 1063.366398] env[61906]: _type = "Task" [ 1063.366398] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.380472] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333677, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.427206] env[61906]: DEBUG nova.network.neutron [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Updating instance_info_cache with network_info: [{"id": "1d0e39a6-9bf4-4034-b7e3-0d500971fc09", "address": "fa:16:3e:2b:07:0a", "network": {"id": "0085e7d2-e26f-4b5d-a282-cf6e0f81786a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1504444833-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00c03ca0de1447da9d47c83f256efdcd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d0e39a6-9b", "ovs_interfaceid": "1d0e39a6-9bf4-4034-b7e3-0d500971fc09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.732886] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a0b802be-be08-49ca-b2bb-c436f0385255 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "256457e7-8d94-46f8-8edb-c5782ac34225" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.609s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.890953] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333677, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.930384] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Releasing lock "refresh_cache-151cc6f4-96c5-4ac3-97d5-73cddfa7b465" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.930700] env[61906]: DEBUG nova.compute.manager [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Instance network_info: |[{"id": "1d0e39a6-9bf4-4034-b7e3-0d500971fc09", "address": "fa:16:3e:2b:07:0a", "network": {"id": "0085e7d2-e26f-4b5d-a282-cf6e0f81786a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1504444833-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00c03ca0de1447da9d47c83f256efdcd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d0e39a6-9b", "ovs_interfaceid": "1d0e39a6-9bf4-4034-b7e3-0d500971fc09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1063.931130] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2b:07:0a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c6934071-bf85-4591-9c7d-55c7ea131262', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1d0e39a6-9bf4-4034-b7e3-0d500971fc09', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1063.938638] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Creating folder: Project (00c03ca0de1447da9d47c83f256efdcd). Parent ref: group-v284713. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1063.941418] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b40e3ab8-6a8e-4486-8fa6-25eb7a2e5e3c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.952872] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Created folder: Project (00c03ca0de1447da9d47c83f256efdcd) in parent group-v284713. [ 1063.953056] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Creating folder: Instances. Parent ref: group-v284835. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1063.953309] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b4c3cc78-eaf4-4f75-a31d-0d31f9f77311 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.956238] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05000ea3-84a1-4ae5-8d91-87b7e825dea1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.963031] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50ee2028-5862-48b8-ad41-d259f410a7f6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.967162] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Created folder: Instances in parent group-v284835. [ 1063.967409] env[61906]: DEBUG oslo.service.loopingcall [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1063.967910] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1063.968146] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-686f00b7-2972-408d-959f-bab1c843bdb8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.007922] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd2b941b-6691-458c-961c-eb96353585ba {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.012729] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1064.012729] env[61906]: value = "task-1333680" [ 1064.012729] env[61906]: _type = "Task" [ 1064.012729] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.019456] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-247a2142-f9f1-4ec0-8382-9efce5cb9ba5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.028535] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333680, 'name': CreateVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.036340] env[61906]: DEBUG nova.compute.provider_tree [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1064.204145] env[61906]: DEBUG nova.compute.manager [req-86cd4c0b-bd38-4b95-9e6c-2a965233b07b req-b3be2331-5cea-48b8-80ea-d5a210e7a502 service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Received event network-changed-1d0e39a6-9bf4-4034-b7e3-0d500971fc09 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1064.204341] env[61906]: DEBUG nova.compute.manager [req-86cd4c0b-bd38-4b95-9e6c-2a965233b07b req-b3be2331-5cea-48b8-80ea-d5a210e7a502 service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Refreshing instance network info cache due to event network-changed-1d0e39a6-9bf4-4034-b7e3-0d500971fc09. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1064.204614] env[61906]: DEBUG oslo_concurrency.lockutils [req-86cd4c0b-bd38-4b95-9e6c-2a965233b07b req-b3be2331-5cea-48b8-80ea-d5a210e7a502 service nova] Acquiring lock "refresh_cache-151cc6f4-96c5-4ac3-97d5-73cddfa7b465" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.204861] env[61906]: DEBUG oslo_concurrency.lockutils [req-86cd4c0b-bd38-4b95-9e6c-2a965233b07b req-b3be2331-5cea-48b8-80ea-d5a210e7a502 service nova] Acquired lock "refresh_cache-151cc6f4-96c5-4ac3-97d5-73cddfa7b465" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.204967] env[61906]: DEBUG nova.network.neutron [req-86cd4c0b-bd38-4b95-9e6c-2a965233b07b req-b3be2331-5cea-48b8-80ea-d5a210e7a502 service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Refreshing network info cache for port 1d0e39a6-9bf4-4034-b7e3-0d500971fc09 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1064.384665] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333677, 'name': ReconfigVM_Task, 'duration_secs': 0.769669} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.384957] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Reconfigured VM instance instance-00000067 to attach disk [datastore1] 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca/216bd7f2-dd02-4d3d-a00c-15f62b3de7ca.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1064.385622] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7676182f-d6b5-4a0f-b7ee-c6542c825fb9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.392327] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1064.392327] env[61906]: value = "task-1333681" [ 1064.392327] env[61906]: _type = "Task" [ 1064.392327] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.400273] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333681, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.525409] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333680, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.539623] env[61906]: DEBUG nova.scheduler.client.report [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1064.877854] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1064.878091] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1064.878244] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Starting heal instance info cache {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1064.878357] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Rebuilding the list of instances to heal {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1064.904131] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333681, 'name': Rename_Task, 'duration_secs': 0.219233} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.905762] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1064.905762] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4c7750dc-2196-4b46-b41a-890582e660e9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.912085] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1064.912085] env[61906]: value = "task-1333682" [ 1064.912085] env[61906]: _type = "Task" [ 1064.912085] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.921512] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333682, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.979084] env[61906]: DEBUG nova.network.neutron [req-86cd4c0b-bd38-4b95-9e6c-2a965233b07b req-b3be2331-5cea-48b8-80ea-d5a210e7a502 service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Updated VIF entry in instance network info cache for port 1d0e39a6-9bf4-4034-b7e3-0d500971fc09. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1064.979627] env[61906]: DEBUG nova.network.neutron [req-86cd4c0b-bd38-4b95-9e6c-2a965233b07b req-b3be2331-5cea-48b8-80ea-d5a210e7a502 service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Updating instance_info_cache with network_info: [{"id": "1d0e39a6-9bf4-4034-b7e3-0d500971fc09", "address": "fa:16:3e:2b:07:0a", "network": {"id": "0085e7d2-e26f-4b5d-a282-cf6e0f81786a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1504444833-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00c03ca0de1447da9d47c83f256efdcd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d0e39a6-9b", "ovs_interfaceid": "1d0e39a6-9bf4-4034-b7e3-0d500971fc09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.028620] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333680, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.047164] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.756s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.073964] env[61906]: INFO nova.scheduler.client.report [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Deleted allocations for instance b3e9ab00-8627-4a57-839e-68be5f794b28 [ 1065.383321] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1065.383530] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1065.412667] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.412827] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquired lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.412979] env[61906]: DEBUG nova.network.neutron [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Forcefully refreshing network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1065.413152] env[61906]: DEBUG nova.objects.instance [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lazy-loading 'info_cache' on Instance uuid 1a335019-915b-4065-b22f-d2f0af728d6f {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1065.423815] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333682, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.485581] env[61906]: DEBUG oslo_concurrency.lockutils [req-86cd4c0b-bd38-4b95-9e6c-2a965233b07b req-b3be2331-5cea-48b8-80ea-d5a210e7a502 service nova] Releasing lock "refresh_cache-151cc6f4-96c5-4ac3-97d5-73cddfa7b465" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.529935] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333680, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.584353] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f6ee43c9-ffee-4aaf-8aff-d25637400ec4 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "b3e9ab00-8627-4a57-839e-68be5f794b28" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.783s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.685447] env[61906]: INFO nova.compute.manager [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Rescuing [ 1065.685715] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "refresh_cache-86be7197-c0d1-49c5-aa30-cf908a506031" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.685893] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquired lock "refresh_cache-86be7197-c0d1-49c5-aa30-cf908a506031" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.686175] env[61906]: DEBUG nova.network.neutron [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1065.926337] env[61906]: DEBUG oslo_vmware.api [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333682, 'name': PowerOnVM_Task, 'duration_secs': 0.691807} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.926711] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1065.926711] env[61906]: INFO nova.compute.manager [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Took 8.74 seconds to spawn the instance on the hypervisor. [ 1065.926789] env[61906]: DEBUG nova.compute.manager [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1065.928042] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0893434-b9c5-4732-a2a2-bc0277cf3573 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.028424] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333680, 'name': CreateVM_Task, 'duration_secs': 1.559929} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.028608] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1066.029318] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.029495] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.029831] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1066.030121] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75c5f219-6c33-4815-8f7e-f5c462694157 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.035172] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1066.035172] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52863fcf-6cde-a858-ba17-ccfd4d2f1791" [ 1066.035172] env[61906]: _type = "Task" [ 1066.035172] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.042506] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52863fcf-6cde-a858-ba17-ccfd4d2f1791, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.229220] env[61906]: DEBUG nova.compute.manager [req-6119652b-9a40-43d4-8c82-c7540b76745a req-bbe013bd-4716-4a94-bb58-4bfe15d897d9 service nova] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Received event network-changed-261ecf70-132a-4fe2-be6b-556a555c86cb {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1066.229430] env[61906]: DEBUG nova.compute.manager [req-6119652b-9a40-43d4-8c82-c7540b76745a req-bbe013bd-4716-4a94-bb58-4bfe15d897d9 service nova] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Refreshing instance network info cache due to event network-changed-261ecf70-132a-4fe2-be6b-556a555c86cb. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1066.229650] env[61906]: DEBUG oslo_concurrency.lockutils [req-6119652b-9a40-43d4-8c82-c7540b76745a req-bbe013bd-4716-4a94-bb58-4bfe15d897d9 service nova] Acquiring lock "refresh_cache-256457e7-8d94-46f8-8edb-c5782ac34225" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.230231] env[61906]: DEBUG oslo_concurrency.lockutils [req-6119652b-9a40-43d4-8c82-c7540b76745a req-bbe013bd-4716-4a94-bb58-4bfe15d897d9 service nova] Acquired lock "refresh_cache-256457e7-8d94-46f8-8edb-c5782ac34225" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.230444] env[61906]: DEBUG nova.network.neutron [req-6119652b-9a40-43d4-8c82-c7540b76745a req-bbe013bd-4716-4a94-bb58-4bfe15d897d9 service nova] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Refreshing network info cache for port 261ecf70-132a-4fe2-be6b-556a555c86cb {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1066.407155] env[61906]: DEBUG nova.network.neutron [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Updating instance_info_cache with network_info: [{"id": "24ab450e-e9c4-4d9a-9c4d-13dee30eb56b", "address": "fa:16:3e:fe:51:59", "network": {"id": "9903d8f6-78ef-4ec6-8386-c031ceb715f6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1583036471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d6fb2749df6498f82aaf7c5529b309f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24ab450e-e9", "ovs_interfaceid": "24ab450e-e9c4-4d9a-9c4d-13dee30eb56b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.446536] env[61906]: INFO nova.compute.manager [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Took 13.58 seconds to build instance. [ 1066.546144] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52863fcf-6cde-a858-ba17-ccfd4d2f1791, 'name': SearchDatastore_Task, 'duration_secs': 0.021191} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.546476] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.546716] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1066.546951] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.547206] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.547410] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1066.547669] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-183fba75-54da-41e6-a350-430c1aa45f4d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.552017] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "7264af54-cd10-4507-8ae3-5bb8a5cd127f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.552243] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "7264af54-cd10-4507-8ae3-5bb8a5cd127f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.552448] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "7264af54-cd10-4507-8ae3-5bb8a5cd127f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.552864] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "7264af54-cd10-4507-8ae3-5bb8a5cd127f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.553261] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "7264af54-cd10-4507-8ae3-5bb8a5cd127f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.555121] env[61906]: INFO nova.compute.manager [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Terminating instance [ 1066.557281] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1066.557539] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1066.558460] env[61906]: DEBUG nova.compute.manager [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1066.558760] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1066.558870] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59329e65-d411-4e38-b3ee-57c472398d19 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.561555] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-240322ac-b142-4f65-bb1c-ed62151c5451 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.567080] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1066.567080] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]520a25e1-2d30-894d-2418-8f017de4224e" [ 1066.567080] env[61906]: _type = "Task" [ 1066.567080] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.569301] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1066.571962] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b8909ed4-6e69-42d2-a818-81b71ff9d437 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.579041] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]520a25e1-2d30-894d-2418-8f017de4224e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.580173] env[61906]: DEBUG oslo_vmware.api [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 1066.580173] env[61906]: value = "task-1333683" [ 1066.580173] env[61906]: _type = "Task" [ 1066.580173] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.589631] env[61906]: DEBUG oslo_vmware.api [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333683, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.910040] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Releasing lock "refresh_cache-86be7197-c0d1-49c5-aa30-cf908a506031" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.933640] env[61906]: DEBUG nova.network.neutron [req-6119652b-9a40-43d4-8c82-c7540b76745a req-bbe013bd-4716-4a94-bb58-4bfe15d897d9 service nova] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Updated VIF entry in instance network info cache for port 261ecf70-132a-4fe2-be6b-556a555c86cb. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1066.934067] env[61906]: DEBUG nova.network.neutron [req-6119652b-9a40-43d4-8c82-c7540b76745a req-bbe013bd-4716-4a94-bb58-4bfe15d897d9 service nova] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Updating instance_info_cache with network_info: [{"id": "261ecf70-132a-4fe2-be6b-556a555c86cb", "address": "fa:16:3e:84:13:a4", "network": {"id": "1c55d0cc-4c30-4b89-8736-3e96c5c9c507", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-245940583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fbdcd2dc0d04e83b9f82a19bdfd0023", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bcd9d2d-25c8-41ad-9a4a-93b9029ba993", "external-id": "nsx-vlan-transportzone-734", "segmentation_id": 734, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap261ecf70-13", "ovs_interfaceid": "261ecf70-132a-4fe2-be6b-556a555c86cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.949804] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5601c1b9-80b8-4ca9-89b4-c81b4eb99f58 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "216bd7f2-dd02-4d3d-a00c-15f62b3de7ca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.098s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.079627] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]520a25e1-2d30-894d-2418-8f017de4224e, 'name': SearchDatastore_Task, 'duration_secs': 0.023156} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.083784] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d44c9f7-27a3-4ef2-89d6-8f7d94690cea {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.091513] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1067.091513] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5211807d-26a9-2de9-6695-73a0572eb47f" [ 1067.091513] env[61906]: _type = "Task" [ 1067.091513] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.095152] env[61906]: DEBUG oslo_vmware.api [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333683, 'name': PowerOffVM_Task, 'duration_secs': 0.45114} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.098158] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1067.098348] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1067.098604] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-78fa76ed-e2cd-4e1c-bf56-11e733afece3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.104677] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5211807d-26a9-2de9-6695-73a0572eb47f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.181626] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1067.181902] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1067.182285] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Deleting the datastore file [datastore2] 7264af54-cd10-4507-8ae3-5bb8a5cd127f {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1067.182560] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-368adcbf-fe08-4ad4-b555-b6a1a12ba3a7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.190573] env[61906]: DEBUG oslo_vmware.api [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for the task: (returnval){ [ 1067.190573] env[61906]: value = "task-1333685" [ 1067.190573] env[61906]: _type = "Task" [ 1067.190573] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.201831] env[61906]: DEBUG oslo_vmware.api [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333685, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.203124] env[61906]: DEBUG nova.network.neutron [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Updating instance_info_cache with network_info: [{"id": "a497b774-60ea-485d-a564-ba7b978b560e", "address": "fa:16:3e:0d:cf:6e", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa497b774-60", "ovs_interfaceid": "a497b774-60ea-485d-a564-ba7b978b560e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.436912] env[61906]: DEBUG oslo_concurrency.lockutils [req-6119652b-9a40-43d4-8c82-c7540b76745a req-bbe013bd-4716-4a94-bb58-4bfe15d897d9 service nova] Releasing lock "refresh_cache-256457e7-8d94-46f8-8edb-c5782ac34225" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.448159] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1067.448444] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-552fb233-3a5e-457b-aeab-8cf5e97a269a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.457211] env[61906]: DEBUG oslo_vmware.api [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1067.457211] env[61906]: value = "task-1333686" [ 1067.457211] env[61906]: _type = "Task" [ 1067.457211] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.465475] env[61906]: DEBUG oslo_vmware.api [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333686, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.606031] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5211807d-26a9-2de9-6695-73a0572eb47f, 'name': SearchDatastore_Task, 'duration_secs': 0.037477} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.606394] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.606619] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 151cc6f4-96c5-4ac3-97d5-73cddfa7b465/151cc6f4-96c5-4ac3-97d5-73cddfa7b465.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1067.606908] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3c6c776a-9849-46cd-b93e-e5fd798c38f2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.614488] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1067.614488] env[61906]: value = "task-1333687" [ 1067.614488] env[61906]: _type = "Task" [ 1067.614488] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.622596] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333687, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.701030] env[61906]: DEBUG oslo_vmware.api [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Task: {'id': task-1333685, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.284471} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.701199] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1067.701436] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1067.701677] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1067.701891] env[61906]: INFO nova.compute.manager [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1067.702199] env[61906]: DEBUG oslo.service.loopingcall [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1067.702475] env[61906]: DEBUG nova.compute.manager [-] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1067.702598] env[61906]: DEBUG nova.network.neutron [-] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1067.705994] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Releasing lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.706238] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Updated the network info_cache for instance {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1067.706445] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.706644] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.706820] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.706988] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.707177] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.707359] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.707498] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61906) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1067.707681] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.966874] env[61906]: DEBUG oslo_vmware.api [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333686, 'name': PowerOffVM_Task, 'duration_secs': 0.262917} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.967267] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1067.968078] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72ed14be-41f3-445c-870b-4de82d078937 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.988473] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a0ef215-4a83-497a-bf29-ec311831add5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.021237] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1068.021756] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e1f144b6-6208-4be3-a141-375e3c2a7868 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.029172] env[61906]: DEBUG oslo_vmware.api [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1068.029172] env[61906]: value = "task-1333688" [ 1068.029172] env[61906]: _type = "Task" [ 1068.029172] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.037519] env[61906]: DEBUG oslo_vmware.api [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333688, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.041118] env[61906]: INFO nova.compute.manager [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Rebuilding instance [ 1068.087677] env[61906]: DEBUG nova.compute.manager [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1068.088643] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd988eae-aea5-498d-ae55-c089025eac03 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.125511] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333687, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.491726} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.125873] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 151cc6f4-96c5-4ac3-97d5-73cddfa7b465/151cc6f4-96c5-4ac3-97d5-73cddfa7b465.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1068.126119] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1068.126376] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-43304726-e4bb-4421-879c-80853dad1b40 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.134128] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1068.134128] env[61906]: value = "task-1333689" [ 1068.134128] env[61906]: _type = "Task" [ 1068.134128] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.144315] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333689, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.210828] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.211127] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.211392] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.211655] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61906) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1068.212922] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b3e432-72ee-429b-9da1-f886eaa1b747 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.223618] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1466e815-ebba-4757-81d6-f0dc8bd90f3c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.244420] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-914aff13-599f-426b-ae19-15f1992f9bd8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.252897] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f41b9e3e-8ac2-4b75-a50b-0f99eb75ebb6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.298072] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180625MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=61906) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1068.298309] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.298598] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.303322] env[61906]: DEBUG nova.compute.manager [req-6f3415c6-c4be-4518-a0a6-fbe1180f4f42 req-c7c6ce61-031a-4da1-8300-88df69f64e6d service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Received event network-vif-deleted-73d0e617-fd5e-4fa2-81c5-710a3beac1a4 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1068.303589] env[61906]: INFO nova.compute.manager [req-6f3415c6-c4be-4518-a0a6-fbe1180f4f42 req-c7c6ce61-031a-4da1-8300-88df69f64e6d service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Neutron deleted interface 73d0e617-fd5e-4fa2-81c5-710a3beac1a4; detaching it from the instance and deleting it from the info cache [ 1068.303857] env[61906]: DEBUG nova.network.neutron [req-6f3415c6-c4be-4518-a0a6-fbe1180f4f42 req-c7c6ce61-031a-4da1-8300-88df69f64e6d service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.422840] env[61906]: DEBUG nova.network.neutron [-] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.540244] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] VM already powered off {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1068.540526] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1068.540798] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.540968] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.541184] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1068.541452] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d81dfd3a-1d0a-4dc6-9291-b2dc8f3da771 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.549895] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1068.550098] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1068.550835] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae06346f-527a-4cbc-b57e-f532d8b54168 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.556672] env[61906]: DEBUG oslo_vmware.api [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1068.556672] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5228572a-b0c1-05d4-8853-8ee6629a1339" [ 1068.556672] env[61906]: _type = "Task" [ 1068.556672] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.565683] env[61906]: DEBUG oslo_vmware.api [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5228572a-b0c1-05d4-8853-8ee6629a1339, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.599856] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1068.600155] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5d7763c3-82d5-44ab-8384-0a2b0def9d67 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.607610] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1068.607610] env[61906]: value = "task-1333690" [ 1068.607610] env[61906]: _type = "Task" [ 1068.607610] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.615835] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333690, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.644129] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333689, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069893} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.644474] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1068.645303] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b92f8d5e-2234-457f-b4ea-c3f0b413e31a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.667782] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] 151cc6f4-96c5-4ac3-97d5-73cddfa7b465/151cc6f4-96c5-4ac3-97d5-73cddfa7b465.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1068.668102] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d37f49d1-090f-4612-a1b5-9fb3705f4664 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.688203] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1068.688203] env[61906]: value = "task-1333691" [ 1068.688203] env[61906]: _type = "Task" [ 1068.688203] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.701332] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333691, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.810520] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c80eb4ab-b5d3-4e22-8015-7c8d26a56bb4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.819453] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6390c0e4-32da-4c34-9ac4-379958c2d6ea {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.846543] env[61906]: DEBUG nova.compute.manager [req-6f3415c6-c4be-4518-a0a6-fbe1180f4f42 req-c7c6ce61-031a-4da1-8300-88df69f64e6d service nova] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Detach interface failed, port_id=73d0e617-fd5e-4fa2-81c5-710a3beac1a4, reason: Instance 7264af54-cd10-4507-8ae3-5bb8a5cd127f could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1068.925533] env[61906]: INFO nova.compute.manager [-] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Took 1.22 seconds to deallocate network for instance. [ 1069.067826] env[61906]: DEBUG oslo_vmware.api [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5228572a-b0c1-05d4-8853-8ee6629a1339, 'name': SearchDatastore_Task, 'duration_secs': 0.012292} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.068629] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f51d6281-bc37-48b9-9cc9-38b661a534c9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.073871] env[61906]: DEBUG oslo_vmware.api [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1069.073871] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5246717a-9270-0eec-7fbd-6965fda1ae0f" [ 1069.073871] env[61906]: _type = "Task" [ 1069.073871] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.081108] env[61906]: DEBUG oslo_vmware.api [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5246717a-9270-0eec-7fbd-6965fda1ae0f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.116558] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333690, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.197796] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333691, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.329051] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 1a335019-915b-4065-b22f-d2f0af728d6f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1069.329051] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 7264af54-cd10-4507-8ae3-5bb8a5cd127f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1069.329051] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 1be91e16-1acb-4f13-9b26-cae496a0efb8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1069.329051] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 86be7197-c0d1-49c5-aa30-cf908a506031 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1069.329292] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1069.329292] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 256457e7-8d94-46f8-8edb-c5782ac34225 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1069.329394] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1069.329508] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 151cc6f4-96c5-4ac3-97d5-73cddfa7b465 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1069.329718] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1069.329877] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1069.426495] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ce3c1ea-6623-4548-8e2b-4d0f6a862875 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.431287] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.434251] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-214575ec-d5d6-455e-a009-aa0e4c8a1b32 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.464267] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87f373a9-92b8-4886-a0be-28040ec4384c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.470868] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6e5a96a-eb72-406e-8572-09b6450a35bd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.484042] env[61906]: DEBUG nova.compute.provider_tree [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1069.588724] env[61906]: DEBUG oslo_vmware.api [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5246717a-9270-0eec-7fbd-6965fda1ae0f, 'name': SearchDatastore_Task, 'duration_secs': 0.045813} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.589197] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1069.589612] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 86be7197-c0d1-49c5-aa30-cf908a506031/a9502563-1f96-46a5-b87e-12c0e9375e6e-rescue.vmdk. {{(pid=61906) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1069.590027] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-74a61104-805d-4d46-b7a8-dda08986bbc9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.598472] env[61906]: DEBUG oslo_vmware.api [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1069.598472] env[61906]: value = "task-1333692" [ 1069.598472] env[61906]: _type = "Task" [ 1069.598472] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.610346] env[61906]: DEBUG oslo_vmware.api [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333692, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.622016] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333690, 'name': PowerOffVM_Task, 'duration_secs': 0.57058} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.622897] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1069.623317] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1069.624609] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-700f5908-d419-4b1a-99b9-4fb946a00a0a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.634656] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1069.635086] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-342c739e-ca78-4531-a369-c499ee10f774 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.700021] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333691, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.710176] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1069.710492] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1069.710722] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Deleting the datastore file [datastore1] 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1069.711045] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d7612b7d-86ac-4317-8425-8d0984a76f15 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.718875] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1069.718875] env[61906]: value = "task-1333694" [ 1069.718875] env[61906]: _type = "Task" [ 1069.718875] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.727360] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333694, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.987547] env[61906]: DEBUG nova.scheduler.client.report [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1070.108609] env[61906]: DEBUG oslo_vmware.api [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333692, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.498781} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.109044] env[61906]: INFO nova.virt.vmwareapi.ds_util [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 86be7197-c0d1-49c5-aa30-cf908a506031/a9502563-1f96-46a5-b87e-12c0e9375e6e-rescue.vmdk. [ 1070.109723] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d889678-00f3-44a1-bcd8-d2cae05ebfcc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.135625] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 86be7197-c0d1-49c5-aa30-cf908a506031/a9502563-1f96-46a5-b87e-12c0e9375e6e-rescue.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1070.135895] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1fa7c40e-4671-4e01-91e3-e7355421bb5a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.155534] env[61906]: DEBUG oslo_vmware.api [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1070.155534] env[61906]: value = "task-1333695" [ 1070.155534] env[61906]: _type = "Task" [ 1070.155534] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.169090] env[61906]: DEBUG oslo_vmware.api [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333695, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.200246] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333691, 'name': ReconfigVM_Task, 'duration_secs': 1.037815} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.200582] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Reconfigured VM instance instance-00000068 to attach disk [datastore1] 151cc6f4-96c5-4ac3-97d5-73cddfa7b465/151cc6f4-96c5-4ac3-97d5-73cddfa7b465.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1070.201295] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9934a0c5-300e-4658-bbfa-6cf711888b2c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.208093] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1070.208093] env[61906]: value = "task-1333696" [ 1070.208093] env[61906]: _type = "Task" [ 1070.208093] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.218020] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333696, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.227586] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333694, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.291964} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.227863] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1070.228099] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1070.228297] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1070.494375] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1070.494596] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.196s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.495016] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.064s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.495293] env[61906]: DEBUG nova.objects.instance [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lazy-loading 'resources' on Instance uuid 7264af54-cd10-4507-8ae3-5bb8a5cd127f {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1070.665793] env[61906]: DEBUG oslo_vmware.api [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333695, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.718492] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333696, 'name': Rename_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.117303] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bacd04b-1c87-435b-b202-d70b262b1ea6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.125089] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-594b6ed2-2882-465f-9227-dfa31b12764f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.155926] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d01229df-fb3a-41be-a423-9fd85d239b4d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.166186] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87051f78-d00d-4256-bc69-3bd782d81e85 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.172669] env[61906]: DEBUG oslo_vmware.api [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333695, 'name': ReconfigVM_Task, 'duration_secs': 0.612886} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.173276] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 86be7197-c0d1-49c5-aa30-cf908a506031/a9502563-1f96-46a5-b87e-12c0e9375e6e-rescue.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1071.174138] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4df2a706-c49c-4d1e-b747-47bebc014b6c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.184091] env[61906]: DEBUG nova.compute.provider_tree [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1071.207393] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b3aa632-035f-4766-8842-ab3a22adbc03 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.226463] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333696, 'name': Rename_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.227639] env[61906]: DEBUG oslo_vmware.api [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1071.227639] env[61906]: value = "task-1333697" [ 1071.227639] env[61906]: _type = "Task" [ 1071.227639] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.235267] env[61906]: DEBUG oslo_vmware.api [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333697, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.263834] env[61906]: DEBUG nova.virt.hardware [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1071.264121] env[61906]: DEBUG nova.virt.hardware [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1071.264294] env[61906]: DEBUG nova.virt.hardware [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1071.264486] env[61906]: DEBUG nova.virt.hardware [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1071.264641] env[61906]: DEBUG nova.virt.hardware [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1071.264795] env[61906]: DEBUG nova.virt.hardware [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1071.265035] env[61906]: DEBUG nova.virt.hardware [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1071.265219] env[61906]: DEBUG nova.virt.hardware [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1071.265398] env[61906]: DEBUG nova.virt.hardware [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1071.265573] env[61906]: DEBUG nova.virt.hardware [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1071.265757] env[61906]: DEBUG nova.virt.hardware [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1071.266775] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f60d6e81-cb74-48e0-abb7-59ea818aede1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.273985] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ef4274f-559a-4be4-9939-0e6d6d148578 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.287444] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e8:71:41', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8716273f-1918-42d4-9166-9788da93cecd', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1071.294849] env[61906]: DEBUG oslo.service.loopingcall [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1071.295098] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1071.295317] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8c33d9db-171f-410f-bf14-478455c8f8e2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.313853] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1071.313853] env[61906]: value = "task-1333698" [ 1071.313853] env[61906]: _type = "Task" [ 1071.313853] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.323212] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333698, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.687983] env[61906]: DEBUG nova.scheduler.client.report [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1071.727159] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333696, 'name': Rename_Task, 'duration_secs': 1.378948} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.727440] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1071.727675] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3becfd82-9aa7-4ac3-af03-a69203fae160 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.737238] env[61906]: DEBUG oslo_vmware.api [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333697, 'name': ReconfigVM_Task, 'duration_secs': 0.235055} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.738407] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1071.738694] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1071.738694] env[61906]: value = "task-1333699" [ 1071.738694] env[61906]: _type = "Task" [ 1071.738694] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.738878] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3c01dae5-aa08-4b32-95d4-cadb5e937b43 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.748454] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333699, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.749503] env[61906]: DEBUG oslo_vmware.api [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1071.749503] env[61906]: value = "task-1333700" [ 1071.749503] env[61906]: _type = "Task" [ 1071.749503] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.757970] env[61906]: DEBUG oslo_vmware.api [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333700, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.824816] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333698, 'name': CreateVM_Task, 'duration_secs': 0.356607} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.825021] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1071.825714] env[61906]: DEBUG oslo_concurrency.lockutils [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.825886] env[61906]: DEBUG oslo_concurrency.lockutils [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.826278] env[61906]: DEBUG oslo_concurrency.lockutils [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1071.826516] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c07622e-b585-4ad7-8474-4c932021bb49 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.830955] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1071.830955] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52e7f364-b11a-61ca-489b-4a8619de54f5" [ 1071.830955] env[61906]: _type = "Task" [ 1071.830955] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.839444] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52e7f364-b11a-61ca-489b-4a8619de54f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.192792] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.698s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.215756] env[61906]: INFO nova.scheduler.client.report [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Deleted allocations for instance 7264af54-cd10-4507-8ae3-5bb8a5cd127f [ 1072.250604] env[61906]: DEBUG oslo_vmware.api [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333699, 'name': PowerOnVM_Task, 'duration_secs': 0.448424} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.253761] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1072.253891] env[61906]: INFO nova.compute.manager [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Took 10.42 seconds to spawn the instance on the hypervisor. [ 1072.254087] env[61906]: DEBUG nova.compute.manager [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1072.255035] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f724ad9-9e53-47c2-99f3-e58ba8065e6c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.261977] env[61906]: DEBUG oslo_vmware.api [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333700, 'name': PowerOnVM_Task, 'duration_secs': 0.429052} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.263425] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1072.268366] env[61906]: DEBUG nova.compute.manager [None req-2f3adb86-9f43-491e-b5db-2641533d1a00 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1072.269083] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e3349ab-f74e-46d8-837f-aa0e67ff1b90 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.341646] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52e7f364-b11a-61ca-489b-4a8619de54f5, 'name': SearchDatastore_Task, 'duration_secs': 0.012764} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.341964] env[61906]: DEBUG oslo_concurrency.lockutils [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.342232] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1072.342478] env[61906]: DEBUG oslo_concurrency.lockutils [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.342630] env[61906]: DEBUG oslo_concurrency.lockutils [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.342813] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1072.343117] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-41dce688-d9ac-4755-b7d6-4cbdc85c9c84 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.350897] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1072.351101] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1072.351782] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfe14c02-5958-4134-b2ee-ee0534cade71 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.357221] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1072.357221] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52d63513-995d-5942-2b8f-dc4c0e0f41ba" [ 1072.357221] env[61906]: _type = "Task" [ 1072.357221] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.364835] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52d63513-995d-5942-2b8f-dc4c0e0f41ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.723433] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5c7194e9-5725-42e8-ae9b-9cbfca8f28d2 tempest-AttachInterfacesTestJSON-344553253 tempest-AttachInterfacesTestJSON-344553253-project-member] Lock "7264af54-cd10-4507-8ae3-5bb8a5cd127f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.171s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.781047] env[61906]: INFO nova.compute.manager [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Took 15.21 seconds to build instance. [ 1072.868115] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52d63513-995d-5942-2b8f-dc4c0e0f41ba, 'name': SearchDatastore_Task, 'duration_secs': 0.008653} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.868966] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a1b1032-3bde-4a19-b44b-36ad4c91bfba {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.873980] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1072.873980] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52676d08-7b26-522a-c6a0-8212c06ed667" [ 1072.873980] env[61906]: _type = "Task" [ 1072.873980] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.881464] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52676d08-7b26-522a-c6a0-8212c06ed667, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.285538] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f5c09249-b3d1-4491-bfa7-6ffa81913711 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.725s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.386943] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52676d08-7b26-522a-c6a0-8212c06ed667, 'name': SearchDatastore_Task, 'duration_secs': 0.009605} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.387288] env[61906]: DEBUG oslo_concurrency.lockutils [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.387559] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca/216bd7f2-dd02-4d3d-a00c-15f62b3de7ca.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1073.387873] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0defe822-5e61-4eb9-aa97-aac4e3f9f1d0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.394938] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1073.394938] env[61906]: value = "task-1333701" [ 1073.394938] env[61906]: _type = "Task" [ 1073.394938] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.403939] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333701, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.644936] env[61906]: DEBUG nova.compute.manager [req-3c9e5564-ae20-44f8-afb9-ecc145d24b24 req-855720fe-ac9c-4f03-a559-4f436d1068f3 service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Received event network-changed-1d0e39a6-9bf4-4034-b7e3-0d500971fc09 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1073.645498] env[61906]: DEBUG nova.compute.manager [req-3c9e5564-ae20-44f8-afb9-ecc145d24b24 req-855720fe-ac9c-4f03-a559-4f436d1068f3 service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Refreshing instance network info cache due to event network-changed-1d0e39a6-9bf4-4034-b7e3-0d500971fc09. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1073.646126] env[61906]: DEBUG oslo_concurrency.lockutils [req-3c9e5564-ae20-44f8-afb9-ecc145d24b24 req-855720fe-ac9c-4f03-a559-4f436d1068f3 service nova] Acquiring lock "refresh_cache-151cc6f4-96c5-4ac3-97d5-73cddfa7b465" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.646585] env[61906]: DEBUG oslo_concurrency.lockutils [req-3c9e5564-ae20-44f8-afb9-ecc145d24b24 req-855720fe-ac9c-4f03-a559-4f436d1068f3 service nova] Acquired lock "refresh_cache-151cc6f4-96c5-4ac3-97d5-73cddfa7b465" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.646830] env[61906]: DEBUG nova.network.neutron [req-3c9e5564-ae20-44f8-afb9-ecc145d24b24 req-855720fe-ac9c-4f03-a559-4f436d1068f3 service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Refreshing network info cache for port 1d0e39a6-9bf4-4034-b7e3-0d500971fc09 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1073.769287] env[61906]: INFO nova.compute.manager [None req-621a97f1-6eea-47d0-9722-cf6e05878527 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Unrescuing [ 1073.769585] env[61906]: DEBUG oslo_concurrency.lockutils [None req-621a97f1-6eea-47d0-9722-cf6e05878527 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "refresh_cache-86be7197-c0d1-49c5-aa30-cf908a506031" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.769749] env[61906]: DEBUG oslo_concurrency.lockutils [None req-621a97f1-6eea-47d0-9722-cf6e05878527 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquired lock "refresh_cache-86be7197-c0d1-49c5-aa30-cf908a506031" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.769928] env[61906]: DEBUG nova.network.neutron [None req-621a97f1-6eea-47d0-9722-cf6e05878527 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1073.906356] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333701, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.492772} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.907258] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca/216bd7f2-dd02-4d3d-a00c-15f62b3de7ca.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1073.907258] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1073.908512] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d32ee3ba-cb52-4bf6-868c-f4fd05c0360a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.918064] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1073.918064] env[61906]: value = "task-1333702" [ 1073.918064] env[61906]: _type = "Task" [ 1073.918064] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.923807] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333702, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.433089] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333702, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071127} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.433089] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1074.433089] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d04466de-bc6a-4aa6-b5f9-650d75e7b852 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.459147] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca/216bd7f2-dd02-4d3d-a00c-15f62b3de7ca.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1074.462953] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23f9804e-e0a6-4911-b545-31f232bf1654 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.479746] env[61906]: DEBUG nova.network.neutron [req-3c9e5564-ae20-44f8-afb9-ecc145d24b24 req-855720fe-ac9c-4f03-a559-4f436d1068f3 service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Updated VIF entry in instance network info cache for port 1d0e39a6-9bf4-4034-b7e3-0d500971fc09. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1074.480351] env[61906]: DEBUG nova.network.neutron [req-3c9e5564-ae20-44f8-afb9-ecc145d24b24 req-855720fe-ac9c-4f03-a559-4f436d1068f3 service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Updating instance_info_cache with network_info: [{"id": "1d0e39a6-9bf4-4034-b7e3-0d500971fc09", "address": "fa:16:3e:2b:07:0a", "network": {"id": "0085e7d2-e26f-4b5d-a282-cf6e0f81786a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1504444833-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.153", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00c03ca0de1447da9d47c83f256efdcd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d0e39a6-9b", "ovs_interfaceid": "1d0e39a6-9bf4-4034-b7e3-0d500971fc09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.489788] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1074.489788] env[61906]: value = "task-1333703" [ 1074.489788] env[61906]: _type = "Task" [ 1074.489788] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.502141] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333703, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.561252] env[61906]: DEBUG nova.network.neutron [None req-621a97f1-6eea-47d0-9722-cf6e05878527 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Updating instance_info_cache with network_info: [{"id": "24ab450e-e9c4-4d9a-9c4d-13dee30eb56b", "address": "fa:16:3e:fe:51:59", "network": {"id": "9903d8f6-78ef-4ec6-8386-c031ceb715f6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1583036471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d6fb2749df6498f82aaf7c5529b309f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24ab450e-e9", "ovs_interfaceid": "24ab450e-e9c4-4d9a-9c4d-13dee30eb56b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.984857] env[61906]: DEBUG oslo_concurrency.lockutils [req-3c9e5564-ae20-44f8-afb9-ecc145d24b24 req-855720fe-ac9c-4f03-a559-4f436d1068f3 service nova] Releasing lock "refresh_cache-151cc6f4-96c5-4ac3-97d5-73cddfa7b465" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1075.000791] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333703, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.064873] env[61906]: DEBUG oslo_concurrency.lockutils [None req-621a97f1-6eea-47d0-9722-cf6e05878527 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Releasing lock "refresh_cache-86be7197-c0d1-49c5-aa30-cf908a506031" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1075.065836] env[61906]: DEBUG nova.objects.instance [None req-621a97f1-6eea-47d0-9722-cf6e05878527 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lazy-loading 'flavor' on Instance uuid 86be7197-c0d1-49c5-aa30-cf908a506031 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1075.499961] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333703, 'name': ReconfigVM_Task, 'duration_secs': 0.546756} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.500320] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca/216bd7f2-dd02-4d3d-a00c-15f62b3de7ca.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1075.500846] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a7e44918-9a1b-445e-8285-271b4a5f795a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.507793] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1075.507793] env[61906]: value = "task-1333704" [ 1075.507793] env[61906]: _type = "Task" [ 1075.507793] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.519515] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333704, 'name': Rename_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.575625] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58e6f801-d91a-4e70-9380-7ad88c8701a4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.598921] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-621a97f1-6eea-47d0-9722-cf6e05878527 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1075.599294] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cc4b573c-4dca-4966-9343-4c5f6a4fab47 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.607550] env[61906]: DEBUG oslo_vmware.api [None req-621a97f1-6eea-47d0-9722-cf6e05878527 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1075.607550] env[61906]: value = "task-1333705" [ 1075.607550] env[61906]: _type = "Task" [ 1075.607550] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.614517] env[61906]: DEBUG oslo_vmware.api [None req-621a97f1-6eea-47d0-9722-cf6e05878527 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333705, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.017014] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333704, 'name': Rename_Task, 'duration_secs': 0.175862} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.017287] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1076.017524] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-84b75d65-9a2d-43f8-b2fb-1258ca068036 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.023786] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1076.023786] env[61906]: value = "task-1333706" [ 1076.023786] env[61906]: _type = "Task" [ 1076.023786] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.032122] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333706, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.117019] env[61906]: DEBUG oslo_vmware.api [None req-621a97f1-6eea-47d0-9722-cf6e05878527 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333705, 'name': PowerOffVM_Task, 'duration_secs': 0.224478} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.117316] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-621a97f1-6eea-47d0-9722-cf6e05878527 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1076.122488] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-621a97f1-6eea-47d0-9722-cf6e05878527 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Reconfiguring VM instance instance-00000064 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1076.122783] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a229294-56af-41dd-98d7-9849e5662207 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.140378] env[61906]: DEBUG oslo_vmware.api [None req-621a97f1-6eea-47d0-9722-cf6e05878527 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1076.140378] env[61906]: value = "task-1333707" [ 1076.140378] env[61906]: _type = "Task" [ 1076.140378] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.147929] env[61906]: DEBUG oslo_vmware.api [None req-621a97f1-6eea-47d0-9722-cf6e05878527 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333707, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.532579] env[61906]: DEBUG oslo_vmware.api [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333706, 'name': PowerOnVM_Task, 'duration_secs': 0.501082} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.532971] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1076.533069] env[61906]: DEBUG nova.compute.manager [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1076.533901] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1935b6d5-e02c-4d02-a7a1-033441ed7fbb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.651297] env[61906]: DEBUG oslo_vmware.api [None req-621a97f1-6eea-47d0-9722-cf6e05878527 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333707, 'name': ReconfigVM_Task, 'duration_secs': 0.261029} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.651601] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-621a97f1-6eea-47d0-9722-cf6e05878527 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Reconfigured VM instance instance-00000064 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1076.651795] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-621a97f1-6eea-47d0-9722-cf6e05878527 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1076.652068] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-812b89ab-e94f-4eab-86de-cb6b06aa6406 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.657785] env[61906]: DEBUG oslo_vmware.api [None req-621a97f1-6eea-47d0-9722-cf6e05878527 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1076.657785] env[61906]: value = "task-1333708" [ 1076.657785] env[61906]: _type = "Task" [ 1076.657785] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.670098] env[61906]: DEBUG oslo_vmware.api [None req-621a97f1-6eea-47d0-9722-cf6e05878527 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333708, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.048994] env[61906]: DEBUG oslo_concurrency.lockutils [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.049170] env[61906]: DEBUG oslo_concurrency.lockutils [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.049258] env[61906]: DEBUG nova.objects.instance [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61906) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1077.167297] env[61906]: DEBUG oslo_vmware.api [None req-621a97f1-6eea-47d0-9722-cf6e05878527 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333708, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.560254] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "216bd7f2-dd02-4d3d-a00c-15f62b3de7ca" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.560646] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "216bd7f2-dd02-4d3d-a00c-15f62b3de7ca" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.560701] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "216bd7f2-dd02-4d3d-a00c-15f62b3de7ca-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.560942] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "216bd7f2-dd02-4d3d-a00c-15f62b3de7ca-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.561179] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "216bd7f2-dd02-4d3d-a00c-15f62b3de7ca-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.563136] env[61906]: INFO nova.compute.manager [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Terminating instance [ 1077.565158] env[61906]: DEBUG nova.compute.manager [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1077.565398] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1077.566315] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1531facc-87f3-47d5-b397-4e2fc78991bb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.574081] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1077.574527] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b76ffa3c-611e-4d0d-98a3-f16e73796871 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.580772] env[61906]: DEBUG oslo_vmware.api [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1077.580772] env[61906]: value = "task-1333709" [ 1077.580772] env[61906]: _type = "Task" [ 1077.580772] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.589575] env[61906]: DEBUG oslo_vmware.api [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333709, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.668050] env[61906]: DEBUG oslo_vmware.api [None req-621a97f1-6eea-47d0-9722-cf6e05878527 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333708, 'name': PowerOnVM_Task, 'duration_secs': 0.848859} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.668355] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-621a97f1-6eea-47d0-9722-cf6e05878527 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1077.668585] env[61906]: DEBUG nova.compute.manager [None req-621a97f1-6eea-47d0-9722-cf6e05878527 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1077.669364] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-485e9f03-de0c-4a45-8125-a0e3d7eff465 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.059111] env[61906]: DEBUG oslo_concurrency.lockutils [None req-386d6a40-061b-4794-9df9-8bbb68dc558e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.091097] env[61906]: DEBUG oslo_vmware.api [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333709, 'name': PowerOffVM_Task, 'duration_secs': 0.221126} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.091387] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1078.091565] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1078.091926] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e3cddde1-2dea-4e63-b1ae-2fa1475fa16e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.158618] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1078.158856] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1078.159061] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Deleting the datastore file [datastore2] 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1078.159343] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3152a1b5-ad08-4e50-adbd-2971e04f2379 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.165680] env[61906]: DEBUG oslo_vmware.api [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1078.165680] env[61906]: value = "task-1333711" [ 1078.165680] env[61906]: _type = "Task" [ 1078.165680] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.173813] env[61906]: DEBUG oslo_vmware.api [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333711, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.675874] env[61906]: DEBUG oslo_vmware.api [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333711, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187309} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.676251] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1078.676345] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1078.676530] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1078.676715] env[61906]: INFO nova.compute.manager [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1078.676972] env[61906]: DEBUG oslo.service.loopingcall [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1078.677199] env[61906]: DEBUG nova.compute.manager [-] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1078.677296] env[61906]: DEBUG nova.network.neutron [-] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1078.930554] env[61906]: DEBUG nova.compute.manager [req-5cbb34ef-3b97-4f8c-b7d9-f076318afb74 req-2b09b894-6d00-4cc0-94ed-deb6cc0a8684 service nova] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Received event network-vif-deleted-8716273f-1918-42d4-9166-9788da93cecd {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1078.930773] env[61906]: INFO nova.compute.manager [req-5cbb34ef-3b97-4f8c-b7d9-f076318afb74 req-2b09b894-6d00-4cc0-94ed-deb6cc0a8684 service nova] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Neutron deleted interface 8716273f-1918-42d4-9166-9788da93cecd; detaching it from the instance and deleting it from the info cache [ 1078.930949] env[61906]: DEBUG nova.network.neutron [req-5cbb34ef-3b97-4f8c-b7d9-f076318afb74 req-2b09b894-6d00-4cc0-94ed-deb6cc0a8684 service nova] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1079.401800] env[61906]: DEBUG nova.network.neutron [-] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1079.433471] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-77fee21d-4d51-4387-96b5-80f2edd5f90d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.443034] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22203ed5-c1de-441c-b23b-5fcb665b072c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.468839] env[61906]: DEBUG nova.compute.manager [req-5cbb34ef-3b97-4f8c-b7d9-f076318afb74 req-2b09b894-6d00-4cc0-94ed-deb6cc0a8684 service nova] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Detach interface failed, port_id=8716273f-1918-42d4-9166-9788da93cecd, reason: Instance 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1079.903696] env[61906]: INFO nova.compute.manager [-] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Took 1.23 seconds to deallocate network for instance. [ 1080.410727] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.411142] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.411472] env[61906]: DEBUG nova.objects.instance [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lazy-loading 'resources' on Instance uuid 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1080.644555] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "d4a6abdd-75ee-4040-8ca8-1863dd40695c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.644834] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "d4a6abdd-75ee-4040-8ca8-1863dd40695c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.014429] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-856fb15c-54c8-421b-9d3e-a891ee34c543 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.022109] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6143b674-d2f1-474c-93e3-4c845fea2116 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.052567] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b6074e-5ff7-41f0-8ccc-c07d4fbdccd0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.059806] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86a09cee-8a34-4053-8b80-a42a58bfa2fe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.072603] env[61906]: DEBUG nova.compute.provider_tree [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1081.147530] env[61906]: DEBUG nova.compute.manager [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1081.575878] env[61906]: DEBUG nova.scheduler.client.report [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1081.665869] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.081133] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.669s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.082973] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.417s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.084557] env[61906]: INFO nova.compute.claims [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1082.099521] env[61906]: INFO nova.scheduler.client.report [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Deleted allocations for instance 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca [ 1082.606287] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1f0b7ca7-9cb3-432b-9c67-5ec28d922022 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "216bd7f2-dd02-4d3d-a00c-15f62b3de7ca" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.046s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.911166] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3159a7e2-b15c-42ba-85ca-b2e73d5a60e4 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "1be91e16-1acb-4f13-9b26-cae496a0efb8" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.911421] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3159a7e2-b15c-42ba-85ca-b2e73d5a60e4 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "1be91e16-1acb-4f13-9b26-cae496a0efb8" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.179523] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-161d2171-7718-46be-8b24-50022b34c504 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.190216] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-337256ea-fcfa-4eb5-83ab-573bd21dced5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.220848] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa442cbb-0b70-496e-9258-d3bec8539074 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.229283] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1ef4667-f1be-4480-8571-f0a993da1c50 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.244490] env[61906]: DEBUG nova.compute.provider_tree [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1083.415456] env[61906]: DEBUG nova.compute.utils [None req-3159a7e2-b15c-42ba-85ca-b2e73d5a60e4 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1083.747922] env[61906]: DEBUG nova.scheduler.client.report [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1083.918290] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3159a7e2-b15c-42ba-85ca-b2e73d5a60e4 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "1be91e16-1acb-4f13-9b26-cae496a0efb8" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.203319] env[61906]: DEBUG nova.compute.manager [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Stashing vm_state: active {{(pid=61906) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1084.252099] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.169s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.252626] env[61906]: DEBUG nova.compute.manager [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1084.723805] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.724163] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.757041] env[61906]: DEBUG nova.compute.utils [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1084.758433] env[61906]: DEBUG nova.compute.manager [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1084.758638] env[61906]: DEBUG nova.network.neutron [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1084.804658] env[61906]: DEBUG nova.policy [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '315d6310be014ebdb4b868fd93a680a0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1d6fb2749df6498f82aaf7c5529b309f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1084.977686] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3159a7e2-b15c-42ba-85ca-b2e73d5a60e4 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "1be91e16-1acb-4f13-9b26-cae496a0efb8" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.977903] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3159a7e2-b15c-42ba-85ca-b2e73d5a60e4 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "1be91e16-1acb-4f13-9b26-cae496a0efb8" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.978225] env[61906]: INFO nova.compute.manager [None req-3159a7e2-b15c-42ba-85ca-b2e73d5a60e4 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Attaching volume 0fe2a5e1-e725-4f64-8ebe-9223495a75f4 to /dev/sdb [ 1085.008513] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b6b5273-d9ce-4038-8e98-23518c1abfb9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.015513] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78c14703-97ae-44a0-96fd-0f007afe4d4e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.028817] env[61906]: DEBUG nova.virt.block_device [None req-3159a7e2-b15c-42ba-85ca-b2e73d5a60e4 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Updating existing volume attachment record: 54809607-6751-41f0-9303-54c9024e2a6f {{(pid=61906) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1085.058023] env[61906]: DEBUG nova.network.neutron [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Successfully created port: 8c2c45ef-6d72-4a38-bce1-9cbc751f3274 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1085.229372] env[61906]: INFO nova.compute.claims [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1085.261707] env[61906]: DEBUG nova.compute.manager [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1085.735548] env[61906]: INFO nova.compute.resource_tracker [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Updating resource usage from migration 0a378672-12b6-4960-b686-6267e9b969d0 [ 1085.844296] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-750aab14-1b41-44ff-b98c-eae41581bcce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.851736] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c15777ed-bd80-43ca-b0f0-1abdb7da7728 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.884019] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d8091b2-ba5a-4c28-b32e-c648d654ee00 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.892332] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f99f9904-ba70-4b44-b369-5ec3832453b5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.906146] env[61906]: DEBUG nova.compute.provider_tree [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1086.274368] env[61906]: DEBUG nova.compute.manager [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1086.299022] env[61906]: DEBUG nova.virt.hardware [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1086.299022] env[61906]: DEBUG nova.virt.hardware [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1086.299216] env[61906]: DEBUG nova.virt.hardware [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1086.299277] env[61906]: DEBUG nova.virt.hardware [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1086.299431] env[61906]: DEBUG nova.virt.hardware [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1086.299585] env[61906]: DEBUG nova.virt.hardware [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1086.299794] env[61906]: DEBUG nova.virt.hardware [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1086.299957] env[61906]: DEBUG nova.virt.hardware [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1086.300144] env[61906]: DEBUG nova.virt.hardware [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1086.300311] env[61906]: DEBUG nova.virt.hardware [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1086.300487] env[61906]: DEBUG nova.virt.hardware [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1086.301344] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-289689c7-31bc-4eb3-9584-b6668a0dfc13 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.309094] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78721557-af7b-4be9-8e81-58239497f0d7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.409235] env[61906]: DEBUG nova.scheduler.client.report [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1086.500888] env[61906]: DEBUG nova.compute.manager [req-e2f64f63-6f67-42fb-aa3f-15a710b5b586 req-616914c4-8f13-41c1-a68c-91ec999cec8d service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Received event network-vif-plugged-8c2c45ef-6d72-4a38-bce1-9cbc751f3274 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1086.501511] env[61906]: DEBUG oslo_concurrency.lockutils [req-e2f64f63-6f67-42fb-aa3f-15a710b5b586 req-616914c4-8f13-41c1-a68c-91ec999cec8d service nova] Acquiring lock "d4a6abdd-75ee-4040-8ca8-1863dd40695c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.501751] env[61906]: DEBUG oslo_concurrency.lockutils [req-e2f64f63-6f67-42fb-aa3f-15a710b5b586 req-616914c4-8f13-41c1-a68c-91ec999cec8d service nova] Lock "d4a6abdd-75ee-4040-8ca8-1863dd40695c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.502079] env[61906]: DEBUG oslo_concurrency.lockutils [req-e2f64f63-6f67-42fb-aa3f-15a710b5b586 req-616914c4-8f13-41c1-a68c-91ec999cec8d service nova] Lock "d4a6abdd-75ee-4040-8ca8-1863dd40695c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.502662] env[61906]: DEBUG nova.compute.manager [req-e2f64f63-6f67-42fb-aa3f-15a710b5b586 req-616914c4-8f13-41c1-a68c-91ec999cec8d service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] No waiting events found dispatching network-vif-plugged-8c2c45ef-6d72-4a38-bce1-9cbc751f3274 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1086.502662] env[61906]: WARNING nova.compute.manager [req-e2f64f63-6f67-42fb-aa3f-15a710b5b586 req-616914c4-8f13-41c1-a68c-91ec999cec8d service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Received unexpected event network-vif-plugged-8c2c45ef-6d72-4a38-bce1-9cbc751f3274 for instance with vm_state building and task_state spawning. [ 1086.598208] env[61906]: DEBUG nova.network.neutron [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Successfully updated port: 8c2c45ef-6d72-4a38-bce1-9cbc751f3274 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1086.913932] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.190s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.914212] env[61906]: INFO nova.compute.manager [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Migrating [ 1087.103496] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "refresh_cache-d4a6abdd-75ee-4040-8ca8-1863dd40695c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.103648] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquired lock "refresh_cache-d4a6abdd-75ee-4040-8ca8-1863dd40695c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.103815] env[61906]: DEBUG nova.network.neutron [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1087.428779] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.429069] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.429195] env[61906]: DEBUG nova.network.neutron [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1087.636808] env[61906]: DEBUG nova.network.neutron [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1087.760546] env[61906]: DEBUG nova.network.neutron [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Updating instance_info_cache with network_info: [{"id": "8c2c45ef-6d72-4a38-bce1-9cbc751f3274", "address": "fa:16:3e:5c:a3:d4", "network": {"id": "9903d8f6-78ef-4ec6-8386-c031ceb715f6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1583036471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d6fb2749df6498f82aaf7c5529b309f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c2c45ef-6d", "ovs_interfaceid": "8c2c45ef-6d72-4a38-bce1-9cbc751f3274", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.127418] env[61906]: DEBUG nova.network.neutron [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Updating instance_info_cache with network_info: [{"id": "a497b774-60ea-485d-a564-ba7b978b560e", "address": "fa:16:3e:0d:cf:6e", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa497b774-60", "ovs_interfaceid": "a497b774-60ea-485d-a564-ba7b978b560e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.264089] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Releasing lock "refresh_cache-d4a6abdd-75ee-4040-8ca8-1863dd40695c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.264250] env[61906]: DEBUG nova.compute.manager [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Instance network_info: |[{"id": "8c2c45ef-6d72-4a38-bce1-9cbc751f3274", "address": "fa:16:3e:5c:a3:d4", "network": {"id": "9903d8f6-78ef-4ec6-8386-c031ceb715f6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1583036471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d6fb2749df6498f82aaf7c5529b309f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c2c45ef-6d", "ovs_interfaceid": "8c2c45ef-6d72-4a38-bce1-9cbc751f3274", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1088.264623] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:a3:d4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '184687d6-125a-4b58-bb5b-fdb404088eda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8c2c45ef-6d72-4a38-bce1-9cbc751f3274', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1088.272168] env[61906]: DEBUG oslo.service.loopingcall [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1088.272717] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1088.272961] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-27c9ce31-1a3c-4ea3-90e6-13d6a8ba575c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.292428] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1088.292428] env[61906]: value = "task-1333714" [ 1088.292428] env[61906]: _type = "Task" [ 1088.292428] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.299744] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333714, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.532456] env[61906]: DEBUG nova.compute.manager [req-75907206-7896-4f50-864c-f7c64a4688af req-9d7e55db-b0d1-49ab-bccd-e09f4d40f4a5 service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Received event network-changed-8c2c45ef-6d72-4a38-bce1-9cbc751f3274 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1088.532456] env[61906]: DEBUG nova.compute.manager [req-75907206-7896-4f50-864c-f7c64a4688af req-9d7e55db-b0d1-49ab-bccd-e09f4d40f4a5 service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Refreshing instance network info cache due to event network-changed-8c2c45ef-6d72-4a38-bce1-9cbc751f3274. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1088.532798] env[61906]: DEBUG oslo_concurrency.lockutils [req-75907206-7896-4f50-864c-f7c64a4688af req-9d7e55db-b0d1-49ab-bccd-e09f4d40f4a5 service nova] Acquiring lock "refresh_cache-d4a6abdd-75ee-4040-8ca8-1863dd40695c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1088.532954] env[61906]: DEBUG oslo_concurrency.lockutils [req-75907206-7896-4f50-864c-f7c64a4688af req-9d7e55db-b0d1-49ab-bccd-e09f4d40f4a5 service nova] Acquired lock "refresh_cache-d4a6abdd-75ee-4040-8ca8-1863dd40695c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.533050] env[61906]: DEBUG nova.network.neutron [req-75907206-7896-4f50-864c-f7c64a4688af req-9d7e55db-b0d1-49ab-bccd-e09f4d40f4a5 service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Refreshing network info cache for port 8c2c45ef-6d72-4a38-bce1-9cbc751f3274 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1088.630489] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Releasing lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.802420] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333714, 'name': CreateVM_Task, 'duration_secs': 0.307253} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.802557] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1088.803229] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1088.803403] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.803748] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1088.804033] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56288450-f6cc-444f-bcaf-d14ed7095bba {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.808506] env[61906]: DEBUG oslo_vmware.api [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1088.808506] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52b6210f-f3c1-6413-73a6-95857bc935c8" [ 1088.808506] env[61906]: _type = "Task" [ 1088.808506] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.815619] env[61906]: DEBUG oslo_vmware.api [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52b6210f-f3c1-6413-73a6-95857bc935c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.230916] env[61906]: DEBUG nova.network.neutron [req-75907206-7896-4f50-864c-f7c64a4688af req-9d7e55db-b0d1-49ab-bccd-e09f4d40f4a5 service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Updated VIF entry in instance network info cache for port 8c2c45ef-6d72-4a38-bce1-9cbc751f3274. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1089.231322] env[61906]: DEBUG nova.network.neutron [req-75907206-7896-4f50-864c-f7c64a4688af req-9d7e55db-b0d1-49ab-bccd-e09f4d40f4a5 service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Updating instance_info_cache with network_info: [{"id": "8c2c45ef-6d72-4a38-bce1-9cbc751f3274", "address": "fa:16:3e:5c:a3:d4", "network": {"id": "9903d8f6-78ef-4ec6-8386-c031ceb715f6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1583036471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d6fb2749df6498f82aaf7c5529b309f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c2c45ef-6d", "ovs_interfaceid": "8c2c45ef-6d72-4a38-bce1-9cbc751f3274", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1089.319300] env[61906]: DEBUG oslo_vmware.api [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52b6210f-f3c1-6413-73a6-95857bc935c8, 'name': SearchDatastore_Task, 'duration_secs': 0.009396} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.319628] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.319879] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1089.320140] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.320297] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.320485] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1089.320767] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-edc3045a-24c7-45a6-bf49-0f88cc5ce77a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.329817] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1089.330014] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1089.330772] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e959f77a-ee5e-438c-b45f-59a9b7a248dd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.337496] env[61906]: DEBUG oslo_vmware.api [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1089.337496] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52e50f40-bb35-493e-707c-ea07edd07e64" [ 1089.337496] env[61906]: _type = "Task" [ 1089.337496] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.345629] env[61906]: DEBUG oslo_vmware.api [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52e50f40-bb35-493e-707c-ea07edd07e64, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.579198] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-3159a7e2-b15c-42ba-85ca-b2e73d5a60e4 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Volume attach. Driver type: vmdk {{(pid=61906) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1089.579460] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-3159a7e2-b15c-42ba-85ca-b2e73d5a60e4 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284839', 'volume_id': '0fe2a5e1-e725-4f64-8ebe-9223495a75f4', 'name': 'volume-0fe2a5e1-e725-4f64-8ebe-9223495a75f4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1be91e16-1acb-4f13-9b26-cae496a0efb8', 'attached_at': '', 'detached_at': '', 'volume_id': '0fe2a5e1-e725-4f64-8ebe-9223495a75f4', 'serial': '0fe2a5e1-e725-4f64-8ebe-9223495a75f4'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1089.580373] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbbc211a-e130-44c7-bcb7-59d7749b2e48 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.597183] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2bc1e7b-ac0a-4b7d-9e26-6b0aaae1dc80 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.623927] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-3159a7e2-b15c-42ba-85ca-b2e73d5a60e4 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] volume-0fe2a5e1-e725-4f64-8ebe-9223495a75f4/volume-0fe2a5e1-e725-4f64-8ebe-9223495a75f4.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1089.624261] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d5f7b83-58e8-43f1-a17d-8d3f3c4dbc42 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.643616] env[61906]: DEBUG oslo_vmware.api [None req-3159a7e2-b15c-42ba-85ca-b2e73d5a60e4 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1089.643616] env[61906]: value = "task-1333715" [ 1089.643616] env[61906]: _type = "Task" [ 1089.643616] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.653768] env[61906]: DEBUG oslo_vmware.api [None req-3159a7e2-b15c-42ba-85ca-b2e73d5a60e4 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333715, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.734598] env[61906]: DEBUG oslo_concurrency.lockutils [req-75907206-7896-4f50-864c-f7c64a4688af req-9d7e55db-b0d1-49ab-bccd-e09f4d40f4a5 service nova] Releasing lock "refresh_cache-d4a6abdd-75ee-4040-8ca8-1863dd40695c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.848462] env[61906]: DEBUG oslo_vmware.api [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52e50f40-bb35-493e-707c-ea07edd07e64, 'name': SearchDatastore_Task, 'duration_secs': 0.010047} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.850165] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-252ca7f7-b6c7-4616-a8c8-52209ff83bf7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.854663] env[61906]: DEBUG oslo_vmware.api [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1089.854663] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]522a0344-7c7d-8805-a3e8-7294d405a066" [ 1089.854663] env[61906]: _type = "Task" [ 1089.854663] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.862168] env[61906]: DEBUG oslo_vmware.api [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]522a0344-7c7d-8805-a3e8-7294d405a066, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.151049] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0dcf475-7eb1-48a5-95c2-052b128ee9e1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.158834] env[61906]: DEBUG oslo_vmware.api [None req-3159a7e2-b15c-42ba-85ca-b2e73d5a60e4 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333715, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.171985] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Updating instance '1a335019-915b-4065-b22f-d2f0af728d6f' progress to 0 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1090.365340] env[61906]: DEBUG oslo_vmware.api [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]522a0344-7c7d-8805-a3e8-7294d405a066, 'name': SearchDatastore_Task, 'duration_secs': 0.010793} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.365573] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.365830] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] d4a6abdd-75ee-4040-8ca8-1863dd40695c/d4a6abdd-75ee-4040-8ca8-1863dd40695c.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1090.366114] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cfa92406-c8a2-47fa-8475-71f81994124b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.372275] env[61906]: DEBUG oslo_vmware.api [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1090.372275] env[61906]: value = "task-1333716" [ 1090.372275] env[61906]: _type = "Task" [ 1090.372275] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.380358] env[61906]: DEBUG oslo_vmware.api [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333716, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.656238] env[61906]: DEBUG oslo_vmware.api [None req-3159a7e2-b15c-42ba-85ca-b2e73d5a60e4 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333715, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.678644] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1090.678954] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b3861305-7667-4ac3-b14f-cb4d10fad7e9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.686237] env[61906]: DEBUG oslo_vmware.api [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1090.686237] env[61906]: value = "task-1333717" [ 1090.686237] env[61906]: _type = "Task" [ 1090.686237] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.694337] env[61906]: DEBUG oslo_vmware.api [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333717, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.881312] env[61906]: DEBUG oslo_vmware.api [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333716, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.160594] env[61906]: DEBUG oslo_vmware.api [None req-3159a7e2-b15c-42ba-85ca-b2e73d5a60e4 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333715, 'name': ReconfigVM_Task, 'duration_secs': 1.357649} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.161108] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-3159a7e2-b15c-42ba-85ca-b2e73d5a60e4 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Reconfigured VM instance instance-00000062 to attach disk [datastore1] volume-0fe2a5e1-e725-4f64-8ebe-9223495a75f4/volume-0fe2a5e1-e725-4f64-8ebe-9223495a75f4.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1091.168801] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3911aeef-43f1-4a27-979d-81f45ae07c13 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.201024] env[61906]: DEBUG oslo_vmware.api [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333717, 'name': PowerOffVM_Task, 'duration_secs': 0.175817} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.202935] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1091.203182] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Updating instance '1a335019-915b-4065-b22f-d2f0af728d6f' progress to 17 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1091.207643] env[61906]: DEBUG oslo_vmware.api [None req-3159a7e2-b15c-42ba-85ca-b2e73d5a60e4 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1091.207643] env[61906]: value = "task-1333718" [ 1091.207643] env[61906]: _type = "Task" [ 1091.207643] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.216238] env[61906]: DEBUG oslo_vmware.api [None req-3159a7e2-b15c-42ba-85ca-b2e73d5a60e4 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333718, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.383109] env[61906]: DEBUG oslo_vmware.api [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333716, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.954725} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.383372] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] d4a6abdd-75ee-4040-8ca8-1863dd40695c/d4a6abdd-75ee-4040-8ca8-1863dd40695c.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1091.383595] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1091.383854] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-12b0924b-10e3-43b4-a114-69f4be6e721f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.390613] env[61906]: DEBUG oslo_vmware.api [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1091.390613] env[61906]: value = "task-1333719" [ 1091.390613] env[61906]: _type = "Task" [ 1091.390613] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.398091] env[61906]: DEBUG oslo_vmware.api [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333719, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.710537] env[61906]: DEBUG nova.virt.hardware [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:23Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1091.710854] env[61906]: DEBUG nova.virt.hardware [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1091.711011] env[61906]: DEBUG nova.virt.hardware [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1091.711252] env[61906]: DEBUG nova.virt.hardware [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1091.711443] env[61906]: DEBUG nova.virt.hardware [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1091.711639] env[61906]: DEBUG nova.virt.hardware [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1091.711906] env[61906]: DEBUG nova.virt.hardware [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1091.712167] env[61906]: DEBUG nova.virt.hardware [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1091.712396] env[61906]: DEBUG nova.virt.hardware [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1091.712602] env[61906]: DEBUG nova.virt.hardware [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1091.712795] env[61906]: DEBUG nova.virt.hardware [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1091.720552] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-308ee655-c528-4dc3-99c6-89cd14e9e277 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.739100] env[61906]: DEBUG oslo_vmware.api [None req-3159a7e2-b15c-42ba-85ca-b2e73d5a60e4 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333718, 'name': ReconfigVM_Task, 'duration_secs': 0.236782} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.740292] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-3159a7e2-b15c-42ba-85ca-b2e73d5a60e4 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284839', 'volume_id': '0fe2a5e1-e725-4f64-8ebe-9223495a75f4', 'name': 'volume-0fe2a5e1-e725-4f64-8ebe-9223495a75f4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1be91e16-1acb-4f13-9b26-cae496a0efb8', 'attached_at': '', 'detached_at': '', 'volume_id': '0fe2a5e1-e725-4f64-8ebe-9223495a75f4', 'serial': '0fe2a5e1-e725-4f64-8ebe-9223495a75f4'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1091.741705] env[61906]: DEBUG oslo_vmware.api [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1091.741705] env[61906]: value = "task-1333720" [ 1091.741705] env[61906]: _type = "Task" [ 1091.741705] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.751015] env[61906]: DEBUG oslo_vmware.api [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333720, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.900031] env[61906]: DEBUG oslo_vmware.api [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333719, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065429} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.900845] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1091.901094] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc4f5d66-5efe-4f27-9faa-d15b81ed1d0d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.923104] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] d4a6abdd-75ee-4040-8ca8-1863dd40695c/d4a6abdd-75ee-4040-8ca8-1863dd40695c.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1091.923451] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a479a677-479e-4616-af4d-86b74e291afe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.943063] env[61906]: DEBUG oslo_vmware.api [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1091.943063] env[61906]: value = "task-1333721" [ 1091.943063] env[61906]: _type = "Task" [ 1091.943063] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.950660] env[61906]: DEBUG oslo_vmware.api [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333721, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.253925] env[61906]: DEBUG oslo_vmware.api [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333720, 'name': ReconfigVM_Task, 'duration_secs': 0.386485} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.254346] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Updating instance '1a335019-915b-4065-b22f-d2f0af728d6f' progress to 33 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1092.454883] env[61906]: DEBUG oslo_vmware.api [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333721, 'name': ReconfigVM_Task, 'duration_secs': 0.28938} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.455153] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Reconfigured VM instance instance-00000069 to attach disk [datastore1] d4a6abdd-75ee-4040-8ca8-1863dd40695c/d4a6abdd-75ee-4040-8ca8-1863dd40695c.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1092.455750] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dc08f30f-668f-4272-b0ba-f2bd7a33a7b0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.461831] env[61906]: DEBUG oslo_vmware.api [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1092.461831] env[61906]: value = "task-1333722" [ 1092.461831] env[61906]: _type = "Task" [ 1092.461831] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.468919] env[61906]: DEBUG oslo_vmware.api [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333722, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.760800] env[61906]: DEBUG nova.virt.hardware [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1092.761087] env[61906]: DEBUG nova.virt.hardware [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1092.761225] env[61906]: DEBUG nova.virt.hardware [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1092.761430] env[61906]: DEBUG nova.virt.hardware [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1092.761577] env[61906]: DEBUG nova.virt.hardware [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1092.761764] env[61906]: DEBUG nova.virt.hardware [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1092.761970] env[61906]: DEBUG nova.virt.hardware [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1092.762343] env[61906]: DEBUG nova.virt.hardware [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1092.762533] env[61906]: DEBUG nova.virt.hardware [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1092.762798] env[61906]: DEBUG nova.virt.hardware [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1092.763013] env[61906]: DEBUG nova.virt.hardware [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1092.768242] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Reconfiguring VM instance instance-0000004e to detach disk 2000 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1092.768541] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb99dc55-f992-419a-a5c8-626afd46bb36 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.781529] env[61906]: DEBUG nova.objects.instance [None req-3159a7e2-b15c-42ba-85ca-b2e73d5a60e4 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lazy-loading 'flavor' on Instance uuid 1be91e16-1acb-4f13-9b26-cae496a0efb8 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1092.789108] env[61906]: DEBUG oslo_vmware.api [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1092.789108] env[61906]: value = "task-1333723" [ 1092.789108] env[61906]: _type = "Task" [ 1092.789108] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.797648] env[61906]: DEBUG oslo_vmware.api [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333723, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.971953] env[61906]: DEBUG oslo_vmware.api [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333722, 'name': Rename_Task, 'duration_secs': 0.128218} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.972270] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1092.972530] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6ac4c20b-de74-42bd-9563-e5ee073eec55 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.982667] env[61906]: DEBUG oslo_vmware.api [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1092.982667] env[61906]: value = "task-1333724" [ 1092.982667] env[61906]: _type = "Task" [ 1092.982667] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.992467] env[61906]: DEBUG oslo_vmware.api [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333724, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.286436] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3159a7e2-b15c-42ba-85ca-b2e73d5a60e4 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "1be91e16-1acb-4f13-9b26-cae496a0efb8" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.308s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.298537] env[61906]: DEBUG oslo_vmware.api [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333723, 'name': ReconfigVM_Task, 'duration_secs': 0.170504} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.298709] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Reconfigured VM instance instance-0000004e to detach disk 2000 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1093.299519] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de4fda56-6546-4bd3-aea4-b41d57037c06 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.322162] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 1a335019-915b-4065-b22f-d2f0af728d6f/1a335019-915b-4065-b22f-d2f0af728d6f.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1093.322719] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d23e7142-70d1-43fe-8469-584a8e72e5b3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.340230] env[61906]: DEBUG oslo_vmware.api [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1093.340230] env[61906]: value = "task-1333725" [ 1093.340230] env[61906]: _type = "Task" [ 1093.340230] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.347749] env[61906]: DEBUG oslo_vmware.api [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333725, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.478842] env[61906]: DEBUG oslo_concurrency.lockutils [None req-84ed8625-886a-4a74-bd27-ed95eb9ed376 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "1be91e16-1acb-4f13-9b26-cae496a0efb8" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.479167] env[61906]: DEBUG oslo_concurrency.lockutils [None req-84ed8625-886a-4a74-bd27-ed95eb9ed376 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "1be91e16-1acb-4f13-9b26-cae496a0efb8" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.493180] env[61906]: DEBUG oslo_vmware.api [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333724, 'name': PowerOnVM_Task, 'duration_secs': 0.435329} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.493438] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1093.493636] env[61906]: INFO nova.compute.manager [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Took 7.22 seconds to spawn the instance on the hypervisor. [ 1093.493818] env[61906]: DEBUG nova.compute.manager [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1093.494632] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cce31a6-30fc-4498-8c19-c21c20afaf64 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.850156] env[61906]: DEBUG oslo_vmware.api [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333725, 'name': ReconfigVM_Task, 'duration_secs': 0.249584} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.850569] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 1a335019-915b-4065-b22f-d2f0af728d6f/1a335019-915b-4065-b22f-d2f0af728d6f.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1093.850736] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Updating instance '1a335019-915b-4065-b22f-d2f0af728d6f' progress to 50 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1093.982384] env[61906]: INFO nova.compute.manager [None req-84ed8625-886a-4a74-bd27-ed95eb9ed376 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Detaching volume 0fe2a5e1-e725-4f64-8ebe-9223495a75f4 [ 1094.013360] env[61906]: INFO nova.virt.block_device [None req-84ed8625-886a-4a74-bd27-ed95eb9ed376 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Attempting to driver detach volume 0fe2a5e1-e725-4f64-8ebe-9223495a75f4 from mountpoint /dev/sdb [ 1094.013660] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-84ed8625-886a-4a74-bd27-ed95eb9ed376 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Volume detach. Driver type: vmdk {{(pid=61906) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1094.013806] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-84ed8625-886a-4a74-bd27-ed95eb9ed376 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284839', 'volume_id': '0fe2a5e1-e725-4f64-8ebe-9223495a75f4', 'name': 'volume-0fe2a5e1-e725-4f64-8ebe-9223495a75f4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1be91e16-1acb-4f13-9b26-cae496a0efb8', 'attached_at': '', 'detached_at': '', 'volume_id': '0fe2a5e1-e725-4f64-8ebe-9223495a75f4', 'serial': '0fe2a5e1-e725-4f64-8ebe-9223495a75f4'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1094.015091] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57280a99-ec5f-4f1c-9c9f-351548d3c169 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.019025] env[61906]: INFO nova.compute.manager [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Took 12.37 seconds to build instance. [ 1094.039450] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b0cea39-90b4-46a8-ba50-226499b31341 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.046978] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-708f41a5-a736-4948-8dd6-9207e1d2e092 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.067451] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abfa9f3e-a7ce-46aa-98e1-042787edf150 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.081990] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-84ed8625-886a-4a74-bd27-ed95eb9ed376 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] The volume has not been displaced from its original location: [datastore1] volume-0fe2a5e1-e725-4f64-8ebe-9223495a75f4/volume-0fe2a5e1-e725-4f64-8ebe-9223495a75f4.vmdk. No consolidation needed. {{(pid=61906) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1094.087396] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-84ed8625-886a-4a74-bd27-ed95eb9ed376 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Reconfiguring VM instance instance-00000062 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1094.087925] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d1e959f-db15-465e-85da-dcbac15714ec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.106197] env[61906]: DEBUG oslo_vmware.api [None req-84ed8625-886a-4a74-bd27-ed95eb9ed376 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1094.106197] env[61906]: value = "task-1333726" [ 1094.106197] env[61906]: _type = "Task" [ 1094.106197] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.113478] env[61906]: DEBUG oslo_vmware.api [None req-84ed8625-886a-4a74-bd27-ed95eb9ed376 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333726, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.357694] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ce1f3ac-4674-475d-bdf3-ecef54e8bd5a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.378382] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9008c48e-c693-4f7f-aa1f-0ae3e162591b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.396110] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Updating instance '1a335019-915b-4065-b22f-d2f0af728d6f' progress to 67 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1094.521364] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2aa0ca29-0ded-4260-8aab-9d4a472b079a tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "d4a6abdd-75ee-4040-8ca8-1863dd40695c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.876s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.616369] env[61906]: DEBUG oslo_vmware.api [None req-84ed8625-886a-4a74-bd27-ed95eb9ed376 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333726, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.775923] env[61906]: DEBUG nova.compute.manager [req-dfc11302-1024-4a70-bb34-8849fe473677 req-7bf1907e-a5c6-4bfc-bd2e-fd7ce2820b6b service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Received event network-changed-8c2c45ef-6d72-4a38-bce1-9cbc751f3274 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1094.776169] env[61906]: DEBUG nova.compute.manager [req-dfc11302-1024-4a70-bb34-8849fe473677 req-7bf1907e-a5c6-4bfc-bd2e-fd7ce2820b6b service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Refreshing instance network info cache due to event network-changed-8c2c45ef-6d72-4a38-bce1-9cbc751f3274. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1094.776398] env[61906]: DEBUG oslo_concurrency.lockutils [req-dfc11302-1024-4a70-bb34-8849fe473677 req-7bf1907e-a5c6-4bfc-bd2e-fd7ce2820b6b service nova] Acquiring lock "refresh_cache-d4a6abdd-75ee-4040-8ca8-1863dd40695c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.776553] env[61906]: DEBUG oslo_concurrency.lockutils [req-dfc11302-1024-4a70-bb34-8849fe473677 req-7bf1907e-a5c6-4bfc-bd2e-fd7ce2820b6b service nova] Acquired lock "refresh_cache-d4a6abdd-75ee-4040-8ca8-1863dd40695c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.776732] env[61906]: DEBUG nova.network.neutron [req-dfc11302-1024-4a70-bb34-8849fe473677 req-7bf1907e-a5c6-4bfc-bd2e-fd7ce2820b6b service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Refreshing network info cache for port 8c2c45ef-6d72-4a38-bce1-9cbc751f3274 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1094.932732] env[61906]: DEBUG nova.network.neutron [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Port a497b774-60ea-485d-a564-ba7b978b560e binding to destination host cpu-1 is already ACTIVE {{(pid=61906) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1095.116913] env[61906]: DEBUG oslo_vmware.api [None req-84ed8625-886a-4a74-bd27-ed95eb9ed376 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333726, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.478131] env[61906]: DEBUG nova.network.neutron [req-dfc11302-1024-4a70-bb34-8849fe473677 req-7bf1907e-a5c6-4bfc-bd2e-fd7ce2820b6b service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Updated VIF entry in instance network info cache for port 8c2c45ef-6d72-4a38-bce1-9cbc751f3274. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1095.478511] env[61906]: DEBUG nova.network.neutron [req-dfc11302-1024-4a70-bb34-8849fe473677 req-7bf1907e-a5c6-4bfc-bd2e-fd7ce2820b6b service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Updating instance_info_cache with network_info: [{"id": "8c2c45ef-6d72-4a38-bce1-9cbc751f3274", "address": "fa:16:3e:5c:a3:d4", "network": {"id": "9903d8f6-78ef-4ec6-8386-c031ceb715f6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1583036471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d6fb2749df6498f82aaf7c5529b309f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c2c45ef-6d", "ovs_interfaceid": "8c2c45ef-6d72-4a38-bce1-9cbc751f3274", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.618787] env[61906]: DEBUG oslo_vmware.api [None req-84ed8625-886a-4a74-bd27-ed95eb9ed376 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333726, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.956546] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "1a335019-915b-4065-b22f-d2f0af728d6f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.956921] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "1a335019-915b-4065-b22f-d2f0af728d6f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.957038] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "1a335019-915b-4065-b22f-d2f0af728d6f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.981322] env[61906]: DEBUG oslo_concurrency.lockutils [req-dfc11302-1024-4a70-bb34-8849fe473677 req-7bf1907e-a5c6-4bfc-bd2e-fd7ce2820b6b service nova] Releasing lock "refresh_cache-d4a6abdd-75ee-4040-8ca8-1863dd40695c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1096.118904] env[61906]: DEBUG oslo_vmware.api [None req-84ed8625-886a-4a74-bd27-ed95eb9ed376 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333726, 'name': ReconfigVM_Task, 'duration_secs': 1.686627} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.119215] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-84ed8625-886a-4a74-bd27-ed95eb9ed376 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Reconfigured VM instance instance-00000062 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1096.124022] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9e572d6-ad88-46c7-bfaa-7749f8c96ccd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.138933] env[61906]: DEBUG oslo_vmware.api [None req-84ed8625-886a-4a74-bd27-ed95eb9ed376 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1096.138933] env[61906]: value = "task-1333727" [ 1096.138933] env[61906]: _type = "Task" [ 1096.138933] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.150565] env[61906]: DEBUG oslo_vmware.api [None req-84ed8625-886a-4a74-bd27-ed95eb9ed376 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333727, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.648910] env[61906]: DEBUG oslo_vmware.api [None req-84ed8625-886a-4a74-bd27-ed95eb9ed376 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333727, 'name': ReconfigVM_Task, 'duration_secs': 0.153589} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.649267] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-84ed8625-886a-4a74-bd27-ed95eb9ed376 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284839', 'volume_id': '0fe2a5e1-e725-4f64-8ebe-9223495a75f4', 'name': 'volume-0fe2a5e1-e725-4f64-8ebe-9223495a75f4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1be91e16-1acb-4f13-9b26-cae496a0efb8', 'attached_at': '', 'detached_at': '', 'volume_id': '0fe2a5e1-e725-4f64-8ebe-9223495a75f4', 'serial': '0fe2a5e1-e725-4f64-8ebe-9223495a75f4'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1096.990693] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.991034] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.991034] env[61906]: DEBUG nova.network.neutron [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1097.192812] env[61906]: DEBUG nova.objects.instance [None req-84ed8625-886a-4a74-bd27-ed95eb9ed376 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lazy-loading 'flavor' on Instance uuid 1be91e16-1acb-4f13-9b26-cae496a0efb8 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1097.697982] env[61906]: DEBUG nova.network.neutron [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Updating instance_info_cache with network_info: [{"id": "a497b774-60ea-485d-a564-ba7b978b560e", "address": "fa:16:3e:0d:cf:6e", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa497b774-60", "ovs_interfaceid": "a497b774-60ea-485d-a564-ba7b978b560e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1098.201623] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Releasing lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1098.205723] env[61906]: DEBUG oslo_concurrency.lockutils [None req-84ed8625-886a-4a74-bd27-ed95eb9ed376 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "1be91e16-1acb-4f13-9b26-cae496a0efb8" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.727s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.726801] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05c6fdba-cef5-422b-9ad7-fc3d19264ebf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.745268] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc4e21bc-e2a2-4d53-bdfd-6695fc6f2587 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.751742] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Updating instance '1a335019-915b-4065-b22f-d2f0af728d6f' progress to 83 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1098.756790] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "1be91e16-1acb-4f13-9b26-cae496a0efb8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.757069] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "1be91e16-1acb-4f13-9b26-cae496a0efb8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.757256] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "1be91e16-1acb-4f13-9b26-cae496a0efb8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.757409] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "1be91e16-1acb-4f13-9b26-cae496a0efb8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.757583] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "1be91e16-1acb-4f13-9b26-cae496a0efb8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.759259] env[61906]: INFO nova.compute.manager [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Terminating instance [ 1098.760818] env[61906]: DEBUG nova.compute.manager [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1098.761051] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1098.761723] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a47266ff-1fde-4017-b421-d1c95246918f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.768089] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1098.768316] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee0c1bf6-4254-473f-bd38-50002ff1a208 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.773346] env[61906]: DEBUG oslo_vmware.api [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1098.773346] env[61906]: value = "task-1333728" [ 1098.773346] env[61906]: _type = "Task" [ 1098.773346] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.780986] env[61906]: DEBUG oslo_vmware.api [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333728, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.257902] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1099.258261] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e0eaa2b9-8f85-4fec-a3e7-e71c6c8e8422 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.265530] env[61906]: DEBUG oslo_vmware.api [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1099.265530] env[61906]: value = "task-1333729" [ 1099.265530] env[61906]: _type = "Task" [ 1099.265530] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.272915] env[61906]: DEBUG oslo_vmware.api [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333729, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.280929] env[61906]: DEBUG oslo_vmware.api [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333728, 'name': PowerOffVM_Task, 'duration_secs': 0.188112} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.281154] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1099.281330] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1099.281608] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-de306db1-d2c4-42c6-96da-e3f3f6b3fdfe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.346770] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1099.347025] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1099.347210] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Deleting the datastore file [datastore1] 1be91e16-1acb-4f13-9b26-cae496a0efb8 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1099.347497] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-99593b06-b2e1-4790-b804-974d3397b900 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.354429] env[61906]: DEBUG oslo_vmware.api [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1099.354429] env[61906]: value = "task-1333731" [ 1099.354429] env[61906]: _type = "Task" [ 1099.354429] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.361619] env[61906]: DEBUG oslo_vmware.api [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333731, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.775602] env[61906]: DEBUG oslo_vmware.api [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333729, 'name': PowerOnVM_Task, 'duration_secs': 0.39986} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.775873] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1099.776073] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f404530a-9d49-4629-b257-b621eeef97b9 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Updating instance '1a335019-915b-4065-b22f-d2f0af728d6f' progress to 100 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1099.864186] env[61906]: DEBUG oslo_vmware.api [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333731, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130026} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.864432] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1099.864602] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1099.864783] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1099.864962] env[61906]: INFO nova.compute.manager [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1099.865252] env[61906]: DEBUG oslo.service.loopingcall [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1099.865461] env[61906]: DEBUG nova.compute.manager [-] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1099.865558] env[61906]: DEBUG nova.network.neutron [-] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1100.299585] env[61906]: DEBUG nova.compute.manager [req-07ba7194-78ef-40e3-a7fe-66520f7f638a req-1fad1ca1-aaa6-4d08-b95a-ae5e7f4639bc service nova] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Received event network-vif-deleted-df3f0533-5c71-429f-9e5b-a39a6eb52ff8 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1100.299860] env[61906]: INFO nova.compute.manager [req-07ba7194-78ef-40e3-a7fe-66520f7f638a req-1fad1ca1-aaa6-4d08-b95a-ae5e7f4639bc service nova] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Neutron deleted interface df3f0533-5c71-429f-9e5b-a39a6eb52ff8; detaching it from the instance and deleting it from the info cache [ 1100.300174] env[61906]: DEBUG nova.network.neutron [req-07ba7194-78ef-40e3-a7fe-66520f7f638a req-1fad1ca1-aaa6-4d08-b95a-ae5e7f4639bc service nova] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1100.778757] env[61906]: DEBUG nova.network.neutron [-] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1100.802507] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9db7e07a-f5c4-43c9-910e-51cb4092c104 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.814758] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1b2967d-8694-411e-acb4-155dc9f7bb41 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.841659] env[61906]: DEBUG nova.compute.manager [req-07ba7194-78ef-40e3-a7fe-66520f7f638a req-1fad1ca1-aaa6-4d08-b95a-ae5e7f4639bc service nova] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Detach interface failed, port_id=df3f0533-5c71-429f-9e5b-a39a6eb52ff8, reason: Instance 1be91e16-1acb-4f13-9b26-cae496a0efb8 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1101.282440] env[61906]: INFO nova.compute.manager [-] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Took 1.42 seconds to deallocate network for instance. [ 1101.795831] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.796231] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.796541] env[61906]: DEBUG nova.objects.instance [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lazy-loading 'resources' on Instance uuid 1be91e16-1acb-4f13-9b26-cae496a0efb8 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1102.405047] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77cf7cd-4109-4931-9409-c40a71df9f95 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.412715] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad7dc8ae-11e2-4f64-ac02-594866e78c07 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.442237] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f865665-545f-44a3-8a22-950c34e510dc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.449131] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f1b9ab0-15cf-40bc-aa39-f1e21b698d45 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.461788] env[61906]: DEBUG nova.compute.provider_tree [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1102.496192] env[61906]: DEBUG oslo_concurrency.lockutils [None req-38dc81f8-a30e-4019-9010-eebf05376306 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "1a335019-915b-4065-b22f-d2f0af728d6f" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.496499] env[61906]: DEBUG oslo_concurrency.lockutils [None req-38dc81f8-a30e-4019-9010-eebf05376306 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "1a335019-915b-4065-b22f-d2f0af728d6f" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.496756] env[61906]: DEBUG nova.compute.manager [None req-38dc81f8-a30e-4019-9010-eebf05376306 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Going to confirm migration 4 {{(pid=61906) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1102.767969] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fd19b3df-aca1-4fb0-b64c-b85b23267df7 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "256457e7-8d94-46f8-8edb-c5782ac34225" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.768263] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fd19b3df-aca1-4fb0-b64c-b85b23267df7 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "256457e7-8d94-46f8-8edb-c5782ac34225" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.965185] env[61906]: DEBUG nova.scheduler.client.report [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1103.032168] env[61906]: DEBUG oslo_concurrency.lockutils [None req-38dc81f8-a30e-4019-9010-eebf05376306 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1103.032390] env[61906]: DEBUG oslo_concurrency.lockutils [None req-38dc81f8-a30e-4019-9010-eebf05376306 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.032570] env[61906]: DEBUG nova.network.neutron [None req-38dc81f8-a30e-4019-9010-eebf05376306 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1103.032754] env[61906]: DEBUG nova.objects.instance [None req-38dc81f8-a30e-4019-9010-eebf05376306 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lazy-loading 'info_cache' on Instance uuid 1a335019-915b-4065-b22f-d2f0af728d6f {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1103.272137] env[61906]: DEBUG nova.compute.utils [None req-fd19b3df-aca1-4fb0-b64c-b85b23267df7 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1103.471100] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.675s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.493024] env[61906]: INFO nova.scheduler.client.report [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Deleted allocations for instance 1be91e16-1acb-4f13-9b26-cae496a0efb8 [ 1103.774855] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fd19b3df-aca1-4fb0-b64c-b85b23267df7 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "256457e7-8d94-46f8-8edb-c5782ac34225" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.000671] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3d7b8868-9199-4af5-b851-b8f4f6322e9e tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "1be91e16-1acb-4f13-9b26-cae496a0efb8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.243s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.435647] env[61906]: DEBUG nova.network.neutron [None req-38dc81f8-a30e-4019-9010-eebf05376306 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Updating instance_info_cache with network_info: [{"id": "a497b774-60ea-485d-a564-ba7b978b560e", "address": "fa:16:3e:0d:cf:6e", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa497b774-60", "ovs_interfaceid": "a497b774-60ea-485d-a564-ba7b978b560e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.842933] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fd19b3df-aca1-4fb0-b64c-b85b23267df7 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "256457e7-8d94-46f8-8edb-c5782ac34225" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.843208] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fd19b3df-aca1-4fb0-b64c-b85b23267df7 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "256457e7-8d94-46f8-8edb-c5782ac34225" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.843466] env[61906]: INFO nova.compute.manager [None req-fd19b3df-aca1-4fb0-b64c-b85b23267df7 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Attaching volume ae96854f-1414-4b09-bb3e-e3014875951a to /dev/sdb [ 1104.878792] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c03543a9-8355-45df-b451-a48ea307966a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.885784] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-805b98af-05e3-4fb2-9ad5-9265bee449da {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.898210] env[61906]: DEBUG nova.virt.block_device [None req-fd19b3df-aca1-4fb0-b64c-b85b23267df7 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Updating existing volume attachment record: 3fbb0a35-caaf-4a4d-bd29-de6e18a0501b {{(pid=61906) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1104.938996] env[61906]: DEBUG oslo_concurrency.lockutils [None req-38dc81f8-a30e-4019-9010-eebf05376306 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Releasing lock "refresh_cache-1a335019-915b-4065-b22f-d2f0af728d6f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.939305] env[61906]: DEBUG nova.objects.instance [None req-38dc81f8-a30e-4019-9010-eebf05376306 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lazy-loading 'migration_context' on Instance uuid 1a335019-915b-4065-b22f-d2f0af728d6f {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1105.384946] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1105.385459] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1105.385725] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1105.385894] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Cleaning up deleted instances {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1105.442872] env[61906]: DEBUG nova.objects.base [None req-38dc81f8-a30e-4019-9010-eebf05376306 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Object Instance<1a335019-915b-4065-b22f-d2f0af728d6f> lazy-loaded attributes: info_cache,migration_context {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1105.443944] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f23c5a1-9564-4e76-bb36-0142c2b7482e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.463922] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b79e4fed-f1fb-4b75-8812-4bf9b9c0a615 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.471839] env[61906]: DEBUG oslo_vmware.api [None req-38dc81f8-a30e-4019-9010-eebf05376306 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1105.471839] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52e55d2e-be80-5945-191d-732240fbca66" [ 1105.471839] env[61906]: _type = "Task" [ 1105.471839] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.481369] env[61906]: DEBUG oslo_vmware.api [None req-38dc81f8-a30e-4019-9010-eebf05376306 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52e55d2e-be80-5945-191d-732240fbca66, 'name': SearchDatastore_Task, 'duration_secs': 0.007193} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.483756] env[61906]: DEBUG oslo_concurrency.lockutils [None req-38dc81f8-a30e-4019-9010-eebf05376306 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.484060] env[61906]: DEBUG oslo_concurrency.lockutils [None req-38dc81f8-a30e-4019-9010-eebf05376306 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.898761] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] There are 40 instances to clean {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1105.899052] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 216bd7f2-dd02-4d3d-a00c-15f62b3de7ca] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1105.956597] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "4146c49d-fd4b-49f6-8b17-76d06fd143af" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.956831] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "4146c49d-fd4b-49f6-8b17-76d06fd143af" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.081195] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b832574-0b18-486f-9d1e-66ba701690c0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.089014] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f236e14c-7428-42bc-a3f1-5266c8445f5e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.119946] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cb6adeb-36a6-4a60-a751-ea70cb88f04e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.127610] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54c91879-c89c-4bb6-92a6-8f3b816077c0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.140611] env[61906]: DEBUG nova.compute.provider_tree [None req-38dc81f8-a30e-4019-9010-eebf05376306 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1106.402794] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 36c6b1a3-be22-4cec-b21c-0445f5988be1] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1106.459318] env[61906]: DEBUG nova.compute.manager [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1106.643649] env[61906]: DEBUG nova.scheduler.client.report [None req-38dc81f8-a30e-4019-9010-eebf05376306 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1106.906360] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 1be91e16-1acb-4f13-9b26-cae496a0efb8] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1106.980405] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.410213] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: ed213a8b-7d7e-47e1-8a4f-d657c13df3ce] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1107.655327] env[61906]: DEBUG oslo_concurrency.lockutils [None req-38dc81f8-a30e-4019-9010-eebf05376306 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.171s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.658281] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.678s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.659818] env[61906]: INFO nova.compute.claims [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1107.913813] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 0fe121ad-a539-4c04-bb65-b524cb3d91a8] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1108.239239] env[61906]: INFO nova.scheduler.client.report [None req-38dc81f8-a30e-4019-9010-eebf05376306 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Deleted allocation for migration 0a378672-12b6-4960-b686-6267e9b969d0 [ 1108.417869] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: df605d77-7766-47b3-bd27-97361ff8ae4d] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1108.745020] env[61906]: DEBUG oslo_concurrency.lockutils [None req-38dc81f8-a30e-4019-9010-eebf05376306 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "1a335019-915b-4065-b22f-d2f0af728d6f" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.248s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.760437] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a54b940b-2eff-4f22-a513-e1ef04be3d3c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.768924] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a68ce86-5bb7-4b54-a524-53635db9f58c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.802092] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-145ff632-f7dd-42eb-a122-357d9b5a40ef {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.808551] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e6f199d-c221-485c-885f-6a276091567c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.821561] env[61906]: DEBUG nova.compute.provider_tree [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1108.921080] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: b3e9ab00-8627-4a57-839e-68be5f794b28] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1109.324642] env[61906]: DEBUG nova.scheduler.client.report [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1109.423954] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: cec87a1e-4176-45bb-a0c7-a594cb399170] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1109.443136] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd19b3df-aca1-4fb0-b64c-b85b23267df7 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Volume attach. Driver type: vmdk {{(pid=61906) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1109.443441] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd19b3df-aca1-4fb0-b64c-b85b23267df7 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284841', 'volume_id': 'ae96854f-1414-4b09-bb3e-e3014875951a', 'name': 'volume-ae96854f-1414-4b09-bb3e-e3014875951a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '256457e7-8d94-46f8-8edb-c5782ac34225', 'attached_at': '', 'detached_at': '', 'volume_id': 'ae96854f-1414-4b09-bb3e-e3014875951a', 'serial': 'ae96854f-1414-4b09-bb3e-e3014875951a'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1109.444393] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-978d248d-01f2-40ca-94c9-21a3f07ae863 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.461218] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b264933-63cc-46d9-983e-ede1ed5c9240 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.485272] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd19b3df-aca1-4fb0-b64c-b85b23267df7 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] volume-ae96854f-1414-4b09-bb3e-e3014875951a/volume-ae96854f-1414-4b09-bb3e-e3014875951a.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1109.485530] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ccc741f-e00f-405a-9480-71a0d5052f37 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.504080] env[61906]: DEBUG oslo_vmware.api [None req-fd19b3df-aca1-4fb0-b64c-b85b23267df7 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 1109.504080] env[61906]: value = "task-1333734" [ 1109.504080] env[61906]: _type = "Task" [ 1109.504080] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.512061] env[61906]: DEBUG oslo_vmware.api [None req-fd19b3df-aca1-4fb0-b64c-b85b23267df7 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333734, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.651919] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "1a335019-915b-4065-b22f-d2f0af728d6f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.652288] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "1a335019-915b-4065-b22f-d2f0af728d6f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.652541] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "1a335019-915b-4065-b22f-d2f0af728d6f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.652733] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "1a335019-915b-4065-b22f-d2f0af728d6f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.652908] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "1a335019-915b-4065-b22f-d2f0af728d6f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.655129] env[61906]: INFO nova.compute.manager [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Terminating instance [ 1109.656900] env[61906]: DEBUG nova.compute.manager [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1109.657108] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1109.657913] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b590c74c-8201-42cf-978c-3874f07a70be {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.665413] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1109.665642] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-84e508d3-0be6-4027-97d2-034cb42cab4d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.671905] env[61906]: DEBUG oslo_vmware.api [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1109.671905] env[61906]: value = "task-1333735" [ 1109.671905] env[61906]: _type = "Task" [ 1109.671905] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.679609] env[61906]: DEBUG oslo_vmware.api [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333735, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.831635] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.172s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.831635] env[61906]: DEBUG nova.compute.manager [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1109.928616] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 34d22e43-882f-45c2-8259-e77424ce0fb6] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1110.013437] env[61906]: DEBUG oslo_vmware.api [None req-fd19b3df-aca1-4fb0-b64c-b85b23267df7 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333734, 'name': ReconfigVM_Task, 'duration_secs': 0.357059} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.013701] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd19b3df-aca1-4fb0-b64c-b85b23267df7 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Reconfigured VM instance instance-00000066 to attach disk [datastore1] volume-ae96854f-1414-4b09-bb3e-e3014875951a/volume-ae96854f-1414-4b09-bb3e-e3014875951a.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1110.018565] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2158adb6-a4af-46a0-a8af-4dec688398c0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.033545] env[61906]: DEBUG oslo_vmware.api [None req-fd19b3df-aca1-4fb0-b64c-b85b23267df7 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 1110.033545] env[61906]: value = "task-1333736" [ 1110.033545] env[61906]: _type = "Task" [ 1110.033545] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.041010] env[61906]: DEBUG oslo_vmware.api [None req-fd19b3df-aca1-4fb0-b64c-b85b23267df7 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333736, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.181854] env[61906]: DEBUG oslo_vmware.api [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333735, 'name': PowerOffVM_Task, 'duration_secs': 0.21803} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.182093] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1110.182269] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1110.182521] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0e6b43ea-9fa9-4646-8b9b-22b5fe6e3a79 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.253709] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1110.253946] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1110.254127] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Deleting the datastore file [datastore1] 1a335019-915b-4065-b22f-d2f0af728d6f {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1110.254469] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-70841f1b-22f8-4e85-98a5-e8e80084ac5b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.261080] env[61906]: DEBUG oslo_vmware.api [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1110.261080] env[61906]: value = "task-1333738" [ 1110.261080] env[61906]: _type = "Task" [ 1110.261080] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.271175] env[61906]: DEBUG oslo_vmware.api [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333738, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.336010] env[61906]: DEBUG nova.compute.utils [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1110.337747] env[61906]: DEBUG nova.compute.manager [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1110.337747] env[61906]: DEBUG nova.network.neutron [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1110.376096] env[61906]: DEBUG nova.policy [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1b017cdae2af4ae8aa069ae23e744797', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '129cdb06b47346b6895df45e73f7b2dd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1110.431653] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 37124c5c-2021-415b-925b-52e23a3d2973] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1110.543888] env[61906]: DEBUG oslo_vmware.api [None req-fd19b3df-aca1-4fb0-b64c-b85b23267df7 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333736, 'name': ReconfigVM_Task, 'duration_secs': 0.144973} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.544233] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd19b3df-aca1-4fb0-b64c-b85b23267df7 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284841', 'volume_id': 'ae96854f-1414-4b09-bb3e-e3014875951a', 'name': 'volume-ae96854f-1414-4b09-bb3e-e3014875951a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '256457e7-8d94-46f8-8edb-c5782ac34225', 'attached_at': '', 'detached_at': '', 'volume_id': 'ae96854f-1414-4b09-bb3e-e3014875951a', 'serial': 'ae96854f-1414-4b09-bb3e-e3014875951a'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1110.645042] env[61906]: DEBUG nova.network.neutron [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Successfully created port: eaae61e5-a80d-4b75-aaf2-149cb19e3409 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1110.771608] env[61906]: DEBUG oslo_vmware.api [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333738, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162077} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.771877] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1110.772085] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1110.772271] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1110.772453] env[61906]: INFO nova.compute.manager [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1110.772709] env[61906]: DEBUG oslo.service.loopingcall [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1110.772902] env[61906]: DEBUG nova.compute.manager [-] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1110.772999] env[61906]: DEBUG nova.network.neutron [-] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1110.841403] env[61906]: DEBUG nova.compute.manager [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1110.935015] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 7264af54-cd10-4507-8ae3-5bb8a5cd127f] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1111.178358] env[61906]: DEBUG nova.compute.manager [req-82cbb5ba-bb7f-4145-99c2-28e3a128440c req-9ca41baa-4a1a-4bfa-b1a7-32e428e5634c service nova] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Received event network-vif-deleted-a497b774-60ea-485d-a564-ba7b978b560e {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1111.178559] env[61906]: INFO nova.compute.manager [req-82cbb5ba-bb7f-4145-99c2-28e3a128440c req-9ca41baa-4a1a-4bfa-b1a7-32e428e5634c service nova] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Neutron deleted interface a497b774-60ea-485d-a564-ba7b978b560e; detaching it from the instance and deleting it from the info cache [ 1111.178756] env[61906]: DEBUG nova.network.neutron [req-82cbb5ba-bb7f-4145-99c2-28e3a128440c req-9ca41baa-4a1a-4bfa-b1a7-32e428e5634c service nova] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.438122] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: a3e3ad00-6921-4072-8cb1-d80302883513] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1111.588164] env[61906]: DEBUG nova.objects.instance [None req-fd19b3df-aca1-4fb0-b64c-b85b23267df7 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lazy-loading 'flavor' on Instance uuid 256457e7-8d94-46f8-8edb-c5782ac34225 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1111.656081] env[61906]: DEBUG nova.network.neutron [-] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.681875] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2d1a05aa-7b8a-4050-8b0d-16bc60c1c36e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.693027] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e054ebbc-e918-4933-9f4e-8d5eeea4c0d2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.723746] env[61906]: DEBUG nova.compute.manager [req-82cbb5ba-bb7f-4145-99c2-28e3a128440c req-9ca41baa-4a1a-4bfa-b1a7-32e428e5634c service nova] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Detach interface failed, port_id=a497b774-60ea-485d-a564-ba7b978b560e, reason: Instance 1a335019-915b-4065-b22f-d2f0af728d6f could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1111.854250] env[61906]: DEBUG nova.compute.manager [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1111.879768] env[61906]: DEBUG nova.virt.hardware [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1111.880455] env[61906]: DEBUG nova.virt.hardware [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1111.882265] env[61906]: DEBUG nova.virt.hardware [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1111.882265] env[61906]: DEBUG nova.virt.hardware [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1111.882265] env[61906]: DEBUG nova.virt.hardware [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1111.882265] env[61906]: DEBUG nova.virt.hardware [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1111.882265] env[61906]: DEBUG nova.virt.hardware [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1111.882265] env[61906]: DEBUG nova.virt.hardware [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1111.882265] env[61906]: DEBUG nova.virt.hardware [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1111.882265] env[61906]: DEBUG nova.virt.hardware [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1111.882524] env[61906]: DEBUG nova.virt.hardware [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1111.883177] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17e4fe87-85f6-4685-8c8c-b5593cf6074e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.891311] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e324461e-01ee-4855-b990-0f2526401340 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.941465] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 458a1669-a62b-4313-874a-e49809d5c034] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1112.026087] env[61906]: DEBUG nova.compute.manager [req-f63d401f-e088-4d78-b6c6-7c0e2db42d46 req-86a6fb92-7862-4f9a-8bce-b4c7b6526d42 service nova] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Received event network-vif-plugged-eaae61e5-a80d-4b75-aaf2-149cb19e3409 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1112.026354] env[61906]: DEBUG oslo_concurrency.lockutils [req-f63d401f-e088-4d78-b6c6-7c0e2db42d46 req-86a6fb92-7862-4f9a-8bce-b4c7b6526d42 service nova] Acquiring lock "4146c49d-fd4b-49f6-8b17-76d06fd143af-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.026580] env[61906]: DEBUG oslo_concurrency.lockutils [req-f63d401f-e088-4d78-b6c6-7c0e2db42d46 req-86a6fb92-7862-4f9a-8bce-b4c7b6526d42 service nova] Lock "4146c49d-fd4b-49f6-8b17-76d06fd143af-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.026807] env[61906]: DEBUG oslo_concurrency.lockutils [req-f63d401f-e088-4d78-b6c6-7c0e2db42d46 req-86a6fb92-7862-4f9a-8bce-b4c7b6526d42 service nova] Lock "4146c49d-fd4b-49f6-8b17-76d06fd143af-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.026919] env[61906]: DEBUG nova.compute.manager [req-f63d401f-e088-4d78-b6c6-7c0e2db42d46 req-86a6fb92-7862-4f9a-8bce-b4c7b6526d42 service nova] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] No waiting events found dispatching network-vif-plugged-eaae61e5-a80d-4b75-aaf2-149cb19e3409 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1112.027113] env[61906]: WARNING nova.compute.manager [req-f63d401f-e088-4d78-b6c6-7c0e2db42d46 req-86a6fb92-7862-4f9a-8bce-b4c7b6526d42 service nova] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Received unexpected event network-vif-plugged-eaae61e5-a80d-4b75-aaf2-149cb19e3409 for instance with vm_state building and task_state spawning. [ 1112.092893] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fd19b3df-aca1-4fb0-b64c-b85b23267df7 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "256457e7-8d94-46f8-8edb-c5782ac34225" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.250s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.158671] env[61906]: INFO nova.compute.manager [-] [instance: 1a335019-915b-4065-b22f-d2f0af728d6f] Took 1.39 seconds to deallocate network for instance. [ 1112.445252] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: f49fb403-fbb4-4e26-8e75-c160d11dea05] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1112.590677] env[61906]: DEBUG nova.network.neutron [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Successfully updated port: eaae61e5-a80d-4b75-aaf2-149cb19e3409 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1112.620462] env[61906]: DEBUG nova.compute.manager [req-a3034917-93d2-4ec0-acc2-34035b636661 req-9e57ac2b-4d36-4c48-91dd-af7d120ae24b service nova] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Received event network-changed-eaae61e5-a80d-4b75-aaf2-149cb19e3409 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1112.620680] env[61906]: DEBUG nova.compute.manager [req-a3034917-93d2-4ec0-acc2-34035b636661 req-9e57ac2b-4d36-4c48-91dd-af7d120ae24b service nova] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Refreshing instance network info cache due to event network-changed-eaae61e5-a80d-4b75-aaf2-149cb19e3409. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1112.620905] env[61906]: DEBUG oslo_concurrency.lockutils [req-a3034917-93d2-4ec0-acc2-34035b636661 req-9e57ac2b-4d36-4c48-91dd-af7d120ae24b service nova] Acquiring lock "refresh_cache-4146c49d-fd4b-49f6-8b17-76d06fd143af" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1112.621069] env[61906]: DEBUG oslo_concurrency.lockutils [req-a3034917-93d2-4ec0-acc2-34035b636661 req-9e57ac2b-4d36-4c48-91dd-af7d120ae24b service nova] Acquired lock "refresh_cache-4146c49d-fd4b-49f6-8b17-76d06fd143af" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.621245] env[61906]: DEBUG nova.network.neutron [req-a3034917-93d2-4ec0-acc2-34035b636661 req-9e57ac2b-4d36-4c48-91dd-af7d120ae24b service nova] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Refreshing network info cache for port eaae61e5-a80d-4b75-aaf2-149cb19e3409 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1112.664511] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.664788] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.664979] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.690964] env[61906]: INFO nova.scheduler.client.report [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Deleted allocations for instance 1a335019-915b-4065-b22f-d2f0af728d6f [ 1112.921579] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6a711b2f-2d8d-422d-837c-1303b0f921d9 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "256457e7-8d94-46f8-8edb-c5782ac34225" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.922178] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6a711b2f-2d8d-422d-837c-1303b0f921d9 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "256457e7-8d94-46f8-8edb-c5782ac34225" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.948394] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 03d3fb1b-0e46-4544-b01d-498a2baf3b45] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1113.093604] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "refresh_cache-4146c49d-fd4b-49f6-8b17-76d06fd143af" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1113.152136] env[61906]: DEBUG nova.network.neutron [req-a3034917-93d2-4ec0-acc2-34035b636661 req-9e57ac2b-4d36-4c48-91dd-af7d120ae24b service nova] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1113.197993] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c11d2067-858c-4d64-8387-85ca0e13a381 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "1a335019-915b-4065-b22f-d2f0af728d6f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.546s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.222138] env[61906]: DEBUG nova.network.neutron [req-a3034917-93d2-4ec0-acc2-34035b636661 req-9e57ac2b-4d36-4c48-91dd-af7d120ae24b service nova] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1113.425343] env[61906]: DEBUG nova.compute.utils [None req-6a711b2f-2d8d-422d-837c-1303b0f921d9 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1113.452490] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: bddfdc7f-09d7-4887-b330-34a596ffa562] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1113.460049] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.460287] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.460466] env[61906]: INFO nova.compute.manager [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Shelving [ 1113.724060] env[61906]: DEBUG oslo_concurrency.lockutils [req-a3034917-93d2-4ec0-acc2-34035b636661 req-9e57ac2b-4d36-4c48-91dd-af7d120ae24b service nova] Releasing lock "refresh_cache-4146c49d-fd4b-49f6-8b17-76d06fd143af" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1113.724506] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquired lock "refresh_cache-4146c49d-fd4b-49f6-8b17-76d06fd143af" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.724662] env[61906]: DEBUG nova.network.neutron [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1113.929317] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6a711b2f-2d8d-422d-837c-1303b0f921d9 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "256457e7-8d94-46f8-8edb-c5782ac34225" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.955306] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 7142e219-6150-4c6f-9632-cbf489035431] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1113.966812] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1113.967083] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-348ccb6e-e10f-47b5-92d7-28406b311f26 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.974166] env[61906]: DEBUG oslo_vmware.api [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1113.974166] env[61906]: value = "task-1333739" [ 1113.974166] env[61906]: _type = "Task" [ 1113.974166] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.982522] env[61906]: DEBUG oslo_vmware.api [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333739, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.254128] env[61906]: DEBUG nova.network.neutron [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1114.380846] env[61906]: DEBUG nova.network.neutron [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Updating instance_info_cache with network_info: [{"id": "eaae61e5-a80d-4b75-aaf2-149cb19e3409", "address": "fa:16:3e:5a:7e:34", "network": {"id": "c3bcef0f-44a7-44f7-8130-df77c0d3ac2d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-330384233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "129cdb06b47346b6895df45e73f7b2dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35e463c7-7d78-4d66-8efd-6127b1f3ee17", "external-id": "nsx-vlan-transportzone-175", "segmentation_id": 175, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeaae61e5-a8", "ovs_interfaceid": "eaae61e5-a80d-4b75-aaf2-149cb19e3409", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1114.458279] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 150bf47d-7da9-4e13-ad00-7a8a25b9504e] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1114.483627] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "e75681a9-c315-4f60-9095-fa60e7e39c22" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.483889] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "e75681a9-c315-4f60-9095-fa60e7e39c22" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.489036] env[61906]: DEBUG oslo_vmware.api [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333739, 'name': PowerOffVM_Task, 'duration_secs': 0.252921} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.489477] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1114.490280] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29f4e460-975e-43c2-bb17-dde8135586f3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.512142] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6a711b2f-2d8d-422d-837c-1303b0f921d9 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "256457e7-8d94-46f8-8edb-c5782ac34225" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.512411] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6a711b2f-2d8d-422d-837c-1303b0f921d9 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "256457e7-8d94-46f8-8edb-c5782ac34225" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.513035] env[61906]: INFO nova.compute.manager [None req-6a711b2f-2d8d-422d-837c-1303b0f921d9 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Attaching volume 4d0201ec-4b0a-4f73-a89a-9f1c8a0b6b50 to /dev/sdc [ 1114.514813] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3e1a622-0d80-4324-b71b-145335199aa3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.549033] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7be049e-a443-4981-a740-11bf6f527be9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.556251] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b94436f5-fd08-41e9-8c9f-7325bdcf3d48 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.569994] env[61906]: DEBUG nova.virt.block_device [None req-6a711b2f-2d8d-422d-837c-1303b0f921d9 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Updating existing volume attachment record: 98fc7c47-8fb9-4386-8cdc-50ca2c6877aa {{(pid=61906) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1114.887727] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Releasing lock "refresh_cache-4146c49d-fd4b-49f6-8b17-76d06fd143af" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1114.888076] env[61906]: DEBUG nova.compute.manager [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Instance network_info: |[{"id": "eaae61e5-a80d-4b75-aaf2-149cb19e3409", "address": "fa:16:3e:5a:7e:34", "network": {"id": "c3bcef0f-44a7-44f7-8130-df77c0d3ac2d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-330384233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "129cdb06b47346b6895df45e73f7b2dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35e463c7-7d78-4d66-8efd-6127b1f3ee17", "external-id": "nsx-vlan-transportzone-175", "segmentation_id": 175, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeaae61e5-a8", "ovs_interfaceid": "eaae61e5-a80d-4b75-aaf2-149cb19e3409", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1114.888551] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5a:7e:34', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35e463c7-7d78-4d66-8efd-6127b1f3ee17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eaae61e5-a80d-4b75-aaf2-149cb19e3409', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1114.896717] env[61906]: DEBUG oslo.service.loopingcall [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1114.896965] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1114.897229] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7af73042-0adb-42a6-8f35-35d6480ae3a3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.917521] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1114.917521] env[61906]: value = "task-1333741" [ 1114.917521] env[61906]: _type = "Task" [ 1114.917521] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.925368] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333741, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.961813] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 0f846f95-2573-4450-9918-a34467d73363] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1114.990633] env[61906]: DEBUG nova.compute.manager [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1115.028520] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Creating Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1115.029296] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-cc90c988-75ae-4e8d-9cb5-69df589a574f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.037286] env[61906]: DEBUG oslo_vmware.api [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1115.037286] env[61906]: value = "task-1333742" [ 1115.037286] env[61906]: _type = "Task" [ 1115.037286] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.047596] env[61906]: DEBUG oslo_vmware.api [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333742, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.427521] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333741, 'name': CreateVM_Task, 'duration_secs': 0.356719} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.427696] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1115.428421] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1115.428602] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.428944] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1115.429225] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fba968bd-9492-451b-92f6-bba5df99165d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.433517] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1115.433517] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5267241d-3792-d9f4-e1c1-eede14b28275" [ 1115.433517] env[61906]: _type = "Task" [ 1115.433517] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.440903] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5267241d-3792-d9f4-e1c1-eede14b28275, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.464775] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 83b59194-b86f-416a-a7fd-18f75551c2c0] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1115.513637] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.513743] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.515365] env[61906]: INFO nova.compute.claims [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1115.547587] env[61906]: DEBUG oslo_vmware.api [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333742, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.943449] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5267241d-3792-d9f4-e1c1-eede14b28275, 'name': SearchDatastore_Task, 'duration_secs': 0.008647} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.943875] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1115.944186] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1115.944496] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1115.944697] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.944927] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1115.945218] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-48a8bf13-819d-4095-af62-7256fce24862 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.952435] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1115.952607] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1115.953271] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2dcb4f3-9dbd-4ebe-82b1-2ed319476087 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.957980] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1115.957980] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52b5be8a-9bfd-783e-e9f7-d52ffb174b29" [ 1115.957980] env[61906]: _type = "Task" [ 1115.957980] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.964866] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52b5be8a-9bfd-783e-e9f7-d52ffb174b29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.968318] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 8dfeea0b-85b2-4c1c-bd92-f2ea56818fe5] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1116.048162] env[61906]: DEBUG oslo_vmware.api [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333742, 'name': CreateSnapshot_Task, 'duration_secs': 0.522155} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.048444] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Created Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1116.049169] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9adad6f3-972c-4649-8020-d81ebddb9ee1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.468511] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52b5be8a-9bfd-783e-e9f7-d52ffb174b29, 'name': SearchDatastore_Task, 'duration_secs': 0.007112} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.469337] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3556ecc8-45d8-4006-b8d2-205901bc9125 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.471669] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 11cb9d89-2304-45a6-b4bb-fcf8417a0518] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1116.476659] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1116.476659] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52e2b207-7c77-d26e-5e29-c869b9827b61" [ 1116.476659] env[61906]: _type = "Task" [ 1116.476659] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.483866] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52e2b207-7c77-d26e-5e29-c869b9827b61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.565858] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Creating linked-clone VM from snapshot {{(pid=61906) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1116.569139] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-93c0c24d-7f7d-46c9-b2ef-66b9318928c8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.576968] env[61906]: DEBUG oslo_vmware.api [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1116.576968] env[61906]: value = "task-1333743" [ 1116.576968] env[61906]: _type = "Task" [ 1116.576968] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.587626] env[61906]: DEBUG oslo_vmware.api [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333743, 'name': CloneVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.622220] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b89610fe-4655-4e77-a13d-6743fd7e0010 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.628855] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69e734b7-2bf4-4a0e-bc44-c9fee9c4039c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.659415] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cab03e6-2c21-40c5-ad59-0f3e024c64c0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.666849] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d46368d-4456-473f-abed-acb88ba08163 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.680662] env[61906]: DEBUG nova.compute.provider_tree [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1116.974833] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: ad2db007-e7aa-4c66-9020-ac4acfcff25a] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1116.986527] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52e2b207-7c77-d26e-5e29-c869b9827b61, 'name': SearchDatastore_Task, 'duration_secs': 0.00846} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.986799] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1116.987379] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 4146c49d-fd4b-49f6-8b17-76d06fd143af/4146c49d-fd4b-49f6-8b17-76d06fd143af.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1116.987379] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8f31cb79-1b7f-453b-a2c5-75e0f934d9c1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.994054] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1116.994054] env[61906]: value = "task-1333745" [ 1116.994054] env[61906]: _type = "Task" [ 1116.994054] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.001703] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333745, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.086815] env[61906]: DEBUG oslo_vmware.api [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333743, 'name': CloneVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.185059] env[61906]: DEBUG nova.scheduler.client.report [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1117.478319] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 0488c6eb-b510-4d33-ab9f-ebfdab63fb3a] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1117.504758] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333745, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.42346} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.505046] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 4146c49d-fd4b-49f6-8b17-76d06fd143af/4146c49d-fd4b-49f6-8b17-76d06fd143af.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1117.505275] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1117.505536] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4b92489b-c396-4259-a58a-9b60dddc123d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.511952] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1117.511952] env[61906]: value = "task-1333746" [ 1117.511952] env[61906]: _type = "Task" [ 1117.511952] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.519429] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333746, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.587903] env[61906]: DEBUG oslo_vmware.api [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333743, 'name': CloneVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.689597] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.176s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.690173] env[61906]: DEBUG nova.compute.manager [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1117.981998] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: e2bf1806-60f8-4d1a-8cc5-fff7ba2c3b15] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1118.023050] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333746, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.023294] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1118.024038] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6775158a-55db-4ab4-b3f5-40c314bbb517 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.045969] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] 4146c49d-fd4b-49f6-8b17-76d06fd143af/4146c49d-fd4b-49f6-8b17-76d06fd143af.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1118.046431] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-54999270-fa03-4440-9bf9-17da63412fbb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.065692] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1118.065692] env[61906]: value = "task-1333747" [ 1118.065692] env[61906]: _type = "Task" [ 1118.065692] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.072830] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333747, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.087053] env[61906]: DEBUG oslo_vmware.api [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333743, 'name': CloneVM_Task, 'duration_secs': 1.035697} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.087053] env[61906]: INFO nova.virt.vmwareapi.vmops [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Created linked-clone VM from snapshot [ 1118.087327] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-642133b0-b705-4e18-bd81-4dadbac4f6cc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.094094] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Uploading image f6304fcd-34fc-4c19-a45b-e5229f8ce75a {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1118.117945] env[61906]: DEBUG oslo_vmware.rw_handles [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1118.117945] env[61906]: value = "vm-284845" [ 1118.117945] env[61906]: _type = "VirtualMachine" [ 1118.117945] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1118.118245] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-14de03e7-1c43-432e-b363-9d0efc1890f5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.124690] env[61906]: DEBUG oslo_vmware.rw_handles [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lease: (returnval){ [ 1118.124690] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]526d45cb-8651-b0b9-96c2-50ed3d755b60" [ 1118.124690] env[61906]: _type = "HttpNfcLease" [ 1118.124690] env[61906]: } obtained for exporting VM: (result){ [ 1118.124690] env[61906]: value = "vm-284845" [ 1118.124690] env[61906]: _type = "VirtualMachine" [ 1118.124690] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1118.124928] env[61906]: DEBUG oslo_vmware.api [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the lease: (returnval){ [ 1118.124928] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]526d45cb-8651-b0b9-96c2-50ed3d755b60" [ 1118.124928] env[61906]: _type = "HttpNfcLease" [ 1118.124928] env[61906]: } to be ready. {{(pid=61906) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1118.131067] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1118.131067] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]526d45cb-8651-b0b9-96c2-50ed3d755b60" [ 1118.131067] env[61906]: _type = "HttpNfcLease" [ 1118.131067] env[61906]: } is initializing. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1118.194924] env[61906]: DEBUG nova.compute.utils [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1118.196291] env[61906]: DEBUG nova.compute.manager [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1118.196457] env[61906]: DEBUG nova.network.neutron [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1118.233534] env[61906]: DEBUG nova.policy [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '27ba0d3b00ce4513a03259970a600190', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b7d3aa88849f4492aa16ff337b8cfdda', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1118.485008] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 0e86bfdf-9002-4a2d-a2ac-af35921b020e] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1118.494756] env[61906]: DEBUG nova.network.neutron [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Successfully created port: 1b29cb2d-b70e-4022-b7c7-fd014b216c1c {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1118.575815] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333747, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.632933] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1118.632933] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]526d45cb-8651-b0b9-96c2-50ed3d755b60" [ 1118.632933] env[61906]: _type = "HttpNfcLease" [ 1118.632933] env[61906]: } is ready. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1118.633254] env[61906]: DEBUG oslo_vmware.rw_handles [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1118.633254] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]526d45cb-8651-b0b9-96c2-50ed3d755b60" [ 1118.633254] env[61906]: _type = "HttpNfcLease" [ 1118.633254] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1118.633979] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3371a811-5f31-4bf0-a2f7-4fcad7e2224b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.641182] env[61906]: DEBUG oslo_vmware.rw_handles [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b1a53b-cc8f-07b4-76c8-6c7e8924607f/disk-0.vmdk from lease info. {{(pid=61906) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1118.641336] env[61906]: DEBUG oslo_vmware.rw_handles [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b1a53b-cc8f-07b4-76c8-6c7e8924607f/disk-0.vmdk for reading. {{(pid=61906) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1118.699462] env[61906]: DEBUG nova.compute.manager [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1118.731683] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1a97d5c8-81de-448a-9024-caf23611b8c0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.988418] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 1bc9f3d4-b539-4841-b5b6-eaefa0d06a72] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1119.076743] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333747, 'name': ReconfigVM_Task, 'duration_secs': 0.734857} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.078386] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Reconfigured VM instance instance-0000006a to attach disk [datastore1] 4146c49d-fd4b-49f6-8b17-76d06fd143af/4146c49d-fd4b-49f6-8b17-76d06fd143af.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1119.079147] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-40268028-4ba6-406d-b589-df37247cbac0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.086328] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1119.086328] env[61906]: value = "task-1333749" [ 1119.086328] env[61906]: _type = "Task" [ 1119.086328] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.098927] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333749, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.116012] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a711b2f-2d8d-422d-837c-1303b0f921d9 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Volume attach. Driver type: vmdk {{(pid=61906) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1119.116273] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a711b2f-2d8d-422d-837c-1303b0f921d9 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284842', 'volume_id': '4d0201ec-4b0a-4f73-a89a-9f1c8a0b6b50', 'name': 'volume-4d0201ec-4b0a-4f73-a89a-9f1c8a0b6b50', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '256457e7-8d94-46f8-8edb-c5782ac34225', 'attached_at': '', 'detached_at': '', 'volume_id': '4d0201ec-4b0a-4f73-a89a-9f1c8a0b6b50', 'serial': '4d0201ec-4b0a-4f73-a89a-9f1c8a0b6b50'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1119.117250] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d48245f7-fc49-42d2-b72b-5e72a6168ab7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.134053] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2710d81-1aa3-4b52-b6ae-2905cba27325 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.163070] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a711b2f-2d8d-422d-837c-1303b0f921d9 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] volume-4d0201ec-4b0a-4f73-a89a-9f1c8a0b6b50/volume-4d0201ec-4b0a-4f73-a89a-9f1c8a0b6b50.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1119.163788] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c15732c9-3a39-4bf7-9724-f63717df6aa7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.183041] env[61906]: DEBUG oslo_vmware.api [None req-6a711b2f-2d8d-422d-837c-1303b0f921d9 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 1119.183041] env[61906]: value = "task-1333750" [ 1119.183041] env[61906]: _type = "Task" [ 1119.183041] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.191385] env[61906]: DEBUG oslo_vmware.api [None req-6a711b2f-2d8d-422d-837c-1303b0f921d9 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333750, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.493072] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: b1159533-c970-49d6-af42-b954b20d92fb] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1119.596946] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333749, 'name': Rename_Task, 'duration_secs': 0.140082} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.597317] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1119.597658] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9891fb1f-165d-43c0-8b5c-7b84bf9df90a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.604346] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1119.604346] env[61906]: value = "task-1333751" [ 1119.604346] env[61906]: _type = "Task" [ 1119.604346] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.612580] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333751, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.693998] env[61906]: DEBUG oslo_vmware.api [None req-6a711b2f-2d8d-422d-837c-1303b0f921d9 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333750, 'name': ReconfigVM_Task, 'duration_secs': 0.472313} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.694630] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a711b2f-2d8d-422d-837c-1303b0f921d9 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Reconfigured VM instance instance-00000066 to attach disk [datastore1] volume-4d0201ec-4b0a-4f73-a89a-9f1c8a0b6b50/volume-4d0201ec-4b0a-4f73-a89a-9f1c8a0b6b50.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1119.699456] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba51b413-3c89-452e-b558-61d985c91da2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.711147] env[61906]: DEBUG nova.compute.manager [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1119.719528] env[61906]: DEBUG oslo_vmware.api [None req-6a711b2f-2d8d-422d-837c-1303b0f921d9 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 1119.719528] env[61906]: value = "task-1333752" [ 1119.719528] env[61906]: _type = "Task" [ 1119.719528] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.728454] env[61906]: DEBUG oslo_vmware.api [None req-6a711b2f-2d8d-422d-837c-1303b0f921d9 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333752, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.738997] env[61906]: DEBUG nova.virt.hardware [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1119.739737] env[61906]: DEBUG nova.virt.hardware [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1119.739737] env[61906]: DEBUG nova.virt.hardware [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1119.739868] env[61906]: DEBUG nova.virt.hardware [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1119.740025] env[61906]: DEBUG nova.virt.hardware [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1119.740190] env[61906]: DEBUG nova.virt.hardware [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1119.740415] env[61906]: DEBUG nova.virt.hardware [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1119.740601] env[61906]: DEBUG nova.virt.hardware [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1119.740816] env[61906]: DEBUG nova.virt.hardware [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1119.741108] env[61906]: DEBUG nova.virt.hardware [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1119.741463] env[61906]: DEBUG nova.virt.hardware [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1119.742326] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10796fe5-b32b-4b1e-a37f-66b5d4ff4c19 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.749852] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e61fd8e-12bf-4acf-a08a-b8ac2483ec58 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.904116] env[61906]: DEBUG nova.compute.manager [req-2cc6b98a-234a-42dc-b814-619727233607 req-f569a38c-828d-4d99-9cb8-982503601388 service nova] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Received event network-vif-plugged-1b29cb2d-b70e-4022-b7c7-fd014b216c1c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1119.904473] env[61906]: DEBUG oslo_concurrency.lockutils [req-2cc6b98a-234a-42dc-b814-619727233607 req-f569a38c-828d-4d99-9cb8-982503601388 service nova] Acquiring lock "e75681a9-c315-4f60-9095-fa60e7e39c22-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.904621] env[61906]: DEBUG oslo_concurrency.lockutils [req-2cc6b98a-234a-42dc-b814-619727233607 req-f569a38c-828d-4d99-9cb8-982503601388 service nova] Lock "e75681a9-c315-4f60-9095-fa60e7e39c22-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.904793] env[61906]: DEBUG oslo_concurrency.lockutils [req-2cc6b98a-234a-42dc-b814-619727233607 req-f569a38c-828d-4d99-9cb8-982503601388 service nova] Lock "e75681a9-c315-4f60-9095-fa60e7e39c22-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.904967] env[61906]: DEBUG nova.compute.manager [req-2cc6b98a-234a-42dc-b814-619727233607 req-f569a38c-828d-4d99-9cb8-982503601388 service nova] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] No waiting events found dispatching network-vif-plugged-1b29cb2d-b70e-4022-b7c7-fd014b216c1c {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1119.905562] env[61906]: WARNING nova.compute.manager [req-2cc6b98a-234a-42dc-b814-619727233607 req-f569a38c-828d-4d99-9cb8-982503601388 service nova] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Received unexpected event network-vif-plugged-1b29cb2d-b70e-4022-b7c7-fd014b216c1c for instance with vm_state building and task_state spawning. [ 1119.996916] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: da493512-d996-4de7-9f47-cadcbc4fbcb9] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1120.049494] env[61906]: DEBUG nova.network.neutron [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Successfully updated port: 1b29cb2d-b70e-4022-b7c7-fd014b216c1c {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1120.114529] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333751, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.230337] env[61906]: DEBUG oslo_vmware.api [None req-6a711b2f-2d8d-422d-837c-1303b0f921d9 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333752, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.500360] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 7504dfa6-bedc-4701-b4fc-60e19e742276] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1120.551942] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "refresh_cache-e75681a9-c315-4f60-9095-fa60e7e39c22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1120.552189] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired lock "refresh_cache-e75681a9-c315-4f60-9095-fa60e7e39c22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1120.552455] env[61906]: DEBUG nova.network.neutron [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1120.615300] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333751, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.730114] env[61906]: DEBUG oslo_vmware.api [None req-6a711b2f-2d8d-422d-837c-1303b0f921d9 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333752, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.003896] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: cc8870d0-eafb-4f73-aa79-b98f51370237] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1121.085595] env[61906]: DEBUG nova.network.neutron [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1121.115340] env[61906]: DEBUG oslo_vmware.api [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333751, 'name': PowerOnVM_Task, 'duration_secs': 1.471501} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.115627] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1121.115837] env[61906]: INFO nova.compute.manager [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Took 9.26 seconds to spawn the instance on the hypervisor. [ 1121.116024] env[61906]: DEBUG nova.compute.manager [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1121.116868] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c593de8-cae1-43e0-9139-792415b220ee {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.231226] env[61906]: DEBUG oslo_vmware.api [None req-6a711b2f-2d8d-422d-837c-1303b0f921d9 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333752, 'name': ReconfigVM_Task, 'duration_secs': 1.051253} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.232031] env[61906]: DEBUG nova.network.neutron [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Updating instance_info_cache with network_info: [{"id": "1b29cb2d-b70e-4022-b7c7-fd014b216c1c", "address": "fa:16:3e:7f:a9:4c", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b29cb2d-b7", "ovs_interfaceid": "1b29cb2d-b70e-4022-b7c7-fd014b216c1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1121.233214] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a711b2f-2d8d-422d-837c-1303b0f921d9 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284842', 'volume_id': '4d0201ec-4b0a-4f73-a89a-9f1c8a0b6b50', 'name': 'volume-4d0201ec-4b0a-4f73-a89a-9f1c8a0b6b50', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '256457e7-8d94-46f8-8edb-c5782ac34225', 'attached_at': '', 'detached_at': '', 'volume_id': '4d0201ec-4b0a-4f73-a89a-9f1c8a0b6b50', 'serial': '4d0201ec-4b0a-4f73-a89a-9f1c8a0b6b50'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1121.507650] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: cb34a0ce-d274-4167-8367-e19809cc9e4d] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1121.636803] env[61906]: INFO nova.compute.manager [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Took 14.67 seconds to build instance. [ 1121.735621] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Releasing lock "refresh_cache-e75681a9-c315-4f60-9095-fa60e7e39c22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1121.735979] env[61906]: DEBUG nova.compute.manager [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Instance network_info: |[{"id": "1b29cb2d-b70e-4022-b7c7-fd014b216c1c", "address": "fa:16:3e:7f:a9:4c", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b29cb2d-b7", "ovs_interfaceid": "1b29cb2d-b70e-4022-b7c7-fd014b216c1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1121.739334] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7f:a9:4c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1b29cb2d-b70e-4022-b7c7-fd014b216c1c', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1121.748714] env[61906]: DEBUG oslo.service.loopingcall [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1121.748976] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1121.749229] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2d330b4c-e2d2-443d-b6a8-0b62e9fc61b7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.770063] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1121.770063] env[61906]: value = "task-1333753" [ 1121.770063] env[61906]: _type = "Task" [ 1121.770063] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.778801] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333753, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.931873] env[61906]: DEBUG nova.compute.manager [req-6ff46f45-aaba-436e-a4aa-5ab70e7764a9 req-32cc477a-805e-452c-b62d-9d57b2a7debd service nova] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Received event network-changed-1b29cb2d-b70e-4022-b7c7-fd014b216c1c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1121.932383] env[61906]: DEBUG nova.compute.manager [req-6ff46f45-aaba-436e-a4aa-5ab70e7764a9 req-32cc477a-805e-452c-b62d-9d57b2a7debd service nova] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Refreshing instance network info cache due to event network-changed-1b29cb2d-b70e-4022-b7c7-fd014b216c1c. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1121.932668] env[61906]: DEBUG oslo_concurrency.lockutils [req-6ff46f45-aaba-436e-a4aa-5ab70e7764a9 req-32cc477a-805e-452c-b62d-9d57b2a7debd service nova] Acquiring lock "refresh_cache-e75681a9-c315-4f60-9095-fa60e7e39c22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1121.933322] env[61906]: DEBUG oslo_concurrency.lockutils [req-6ff46f45-aaba-436e-a4aa-5ab70e7764a9 req-32cc477a-805e-452c-b62d-9d57b2a7debd service nova] Acquired lock "refresh_cache-e75681a9-c315-4f60-9095-fa60e7e39c22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.933557] env[61906]: DEBUG nova.network.neutron [req-6ff46f45-aaba-436e-a4aa-5ab70e7764a9 req-32cc477a-805e-452c-b62d-9d57b2a7debd service nova] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Refreshing network info cache for port 1b29cb2d-b70e-4022-b7c7-fd014b216c1c {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1122.010610] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 75b9738e-e9b6-435a-88bb-851982828d36] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1122.138757] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4e78b82a-6dfc-4411-a361-08217b479b88 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "4146c49d-fd4b-49f6-8b17-76d06fd143af" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.182s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.271809] env[61906]: DEBUG nova.objects.instance [None req-6a711b2f-2d8d-422d-837c-1303b0f921d9 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lazy-loading 'flavor' on Instance uuid 256457e7-8d94-46f8-8edb-c5782ac34225 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1122.282713] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333753, 'name': CreateVM_Task, 'duration_secs': 0.353036} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.282871] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1122.285899] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1122.285899] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.285899] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1122.285899] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-896654ab-d8b0-4560-a5a7-132971c62406 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.290440] env[61906]: DEBUG oslo_vmware.api [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1122.290440] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]522e5ba0-efe0-5c8d-973d-3b4a42057cb9" [ 1122.290440] env[61906]: _type = "Task" [ 1122.290440] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.298655] env[61906]: DEBUG oslo_vmware.api [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]522e5ba0-efe0-5c8d-973d-3b4a42057cb9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.514556] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 3e0d4b98-a060-48fe-8e6b-f3d7144ae7e8] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1122.554442] env[61906]: DEBUG nova.compute.manager [req-7b7da7ac-5610-4713-a10f-7eb6487cd8e7 req-ec0aee48-b1da-487d-a14e-bf20a05dfd25 service nova] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Received event network-changed-eaae61e5-a80d-4b75-aaf2-149cb19e3409 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1122.555208] env[61906]: DEBUG nova.compute.manager [req-7b7da7ac-5610-4713-a10f-7eb6487cd8e7 req-ec0aee48-b1da-487d-a14e-bf20a05dfd25 service nova] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Refreshing instance network info cache due to event network-changed-eaae61e5-a80d-4b75-aaf2-149cb19e3409. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1122.555208] env[61906]: DEBUG oslo_concurrency.lockutils [req-7b7da7ac-5610-4713-a10f-7eb6487cd8e7 req-ec0aee48-b1da-487d-a14e-bf20a05dfd25 service nova] Acquiring lock "refresh_cache-4146c49d-fd4b-49f6-8b17-76d06fd143af" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1122.555208] env[61906]: DEBUG oslo_concurrency.lockutils [req-7b7da7ac-5610-4713-a10f-7eb6487cd8e7 req-ec0aee48-b1da-487d-a14e-bf20a05dfd25 service nova] Acquired lock "refresh_cache-4146c49d-fd4b-49f6-8b17-76d06fd143af" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.555372] env[61906]: DEBUG nova.network.neutron [req-7b7da7ac-5610-4713-a10f-7eb6487cd8e7 req-ec0aee48-b1da-487d-a14e-bf20a05dfd25 service nova] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Refreshing network info cache for port eaae61e5-a80d-4b75-aaf2-149cb19e3409 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1122.694865] env[61906]: DEBUG nova.network.neutron [req-6ff46f45-aaba-436e-a4aa-5ab70e7764a9 req-32cc477a-805e-452c-b62d-9d57b2a7debd service nova] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Updated VIF entry in instance network info cache for port 1b29cb2d-b70e-4022-b7c7-fd014b216c1c. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1122.695266] env[61906]: DEBUG nova.network.neutron [req-6ff46f45-aaba-436e-a4aa-5ab70e7764a9 req-32cc477a-805e-452c-b62d-9d57b2a7debd service nova] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Updating instance_info_cache with network_info: [{"id": "1b29cb2d-b70e-4022-b7c7-fd014b216c1c", "address": "fa:16:3e:7f:a9:4c", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b29cb2d-b7", "ovs_interfaceid": "1b29cb2d-b70e-4022-b7c7-fd014b216c1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1122.781746] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6a711b2f-2d8d-422d-837c-1303b0f921d9 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "256457e7-8d94-46f8-8edb-c5782ac34225" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.267s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.800985] env[61906]: DEBUG oslo_vmware.api [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]522e5ba0-efe0-5c8d-973d-3b4a42057cb9, 'name': SearchDatastore_Task, 'duration_secs': 0.014945} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.801331] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1122.801963] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1122.801963] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1122.801963] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.802158] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1122.803027] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-277c528a-5fff-4200-bfbd-4b8731e7fb6b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.812245] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1122.812458] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1122.813194] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c8b40f9-7a2a-4e3c-bb78-3a5ec5630489 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.819478] env[61906]: DEBUG oslo_vmware.api [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1122.819478] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52d7ea13-dbcb-9c57-fe8b-3d53ef03471d" [ 1122.819478] env[61906]: _type = "Task" [ 1122.819478] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.828706] env[61906]: DEBUG oslo_vmware.api [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52d7ea13-dbcb-9c57-fe8b-3d53ef03471d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.018659] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: a3a0af30-f0b3-4d38-ba38-a9ddf978d3f8] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1123.198014] env[61906]: DEBUG oslo_concurrency.lockutils [req-6ff46f45-aaba-436e-a4aa-5ab70e7764a9 req-32cc477a-805e-452c-b62d-9d57b2a7debd service nova] Releasing lock "refresh_cache-e75681a9-c315-4f60-9095-fa60e7e39c22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1123.277855] env[61906]: DEBUG nova.network.neutron [req-7b7da7ac-5610-4713-a10f-7eb6487cd8e7 req-ec0aee48-b1da-487d-a14e-bf20a05dfd25 service nova] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Updated VIF entry in instance network info cache for port eaae61e5-a80d-4b75-aaf2-149cb19e3409. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1123.278271] env[61906]: DEBUG nova.network.neutron [req-7b7da7ac-5610-4713-a10f-7eb6487cd8e7 req-ec0aee48-b1da-487d-a14e-bf20a05dfd25 service nova] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Updating instance_info_cache with network_info: [{"id": "eaae61e5-a80d-4b75-aaf2-149cb19e3409", "address": "fa:16:3e:5a:7e:34", "network": {"id": "c3bcef0f-44a7-44f7-8130-df77c0d3ac2d", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-330384233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.167", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "129cdb06b47346b6895df45e73f7b2dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35e463c7-7d78-4d66-8efd-6127b1f3ee17", "external-id": "nsx-vlan-transportzone-175", "segmentation_id": 175, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeaae61e5-a8", "ovs_interfaceid": "eaae61e5-a80d-4b75-aaf2-149cb19e3409", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1123.330105] env[61906]: DEBUG oslo_vmware.api [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52d7ea13-dbcb-9c57-fe8b-3d53ef03471d, 'name': SearchDatastore_Task, 'duration_secs': 0.015442} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.331045] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3a04ed8-2fbc-4e09-a08b-8cef9ad4df14 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.336788] env[61906]: DEBUG oslo_vmware.api [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1123.336788] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52c32870-7f07-c9a5-a4ea-cf12d63c5d86" [ 1123.336788] env[61906]: _type = "Task" [ 1123.336788] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.344479] env[61906]: DEBUG oslo_vmware.api [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52c32870-7f07-c9a5-a4ea-cf12d63c5d86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.521907] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: d60575a6-af13-4d81-95d1-081d0f3e2ce6] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1123.601552] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dbad5f80-74b1-4de3-a428-53aa48cdd965 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "256457e7-8d94-46f8-8edb-c5782ac34225" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.601919] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dbad5f80-74b1-4de3-a428-53aa48cdd965 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "256457e7-8d94-46f8-8edb-c5782ac34225" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.780788] env[61906]: DEBUG oslo_concurrency.lockutils [req-7b7da7ac-5610-4713-a10f-7eb6487cd8e7 req-ec0aee48-b1da-487d-a14e-bf20a05dfd25 service nova] Releasing lock "refresh_cache-4146c49d-fd4b-49f6-8b17-76d06fd143af" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1123.849190] env[61906]: DEBUG oslo_vmware.api [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52c32870-7f07-c9a5-a4ea-cf12d63c5d86, 'name': SearchDatastore_Task, 'duration_secs': 0.016978} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.849581] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1123.849888] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] e75681a9-c315-4f60-9095-fa60e7e39c22/e75681a9-c315-4f60-9095-fa60e7e39c22.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1123.850235] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-28d429a3-484d-4eeb-9973-59b9d24120db {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.858169] env[61906]: DEBUG oslo_vmware.api [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1123.858169] env[61906]: value = "task-1333754" [ 1123.858169] env[61906]: _type = "Task" [ 1123.858169] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.866576] env[61906]: DEBUG oslo_vmware.api [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333754, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.025118] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 80a6a0b8-17c3-46e8-8861-6d1759a19111] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1124.106485] env[61906]: INFO nova.compute.manager [None req-dbad5f80-74b1-4de3-a428-53aa48cdd965 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Detaching volume ae96854f-1414-4b09-bb3e-e3014875951a [ 1124.145349] env[61906]: INFO nova.virt.block_device [None req-dbad5f80-74b1-4de3-a428-53aa48cdd965 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Attempting to driver detach volume ae96854f-1414-4b09-bb3e-e3014875951a from mountpoint /dev/sdb [ 1124.145835] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbad5f80-74b1-4de3-a428-53aa48cdd965 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Volume detach. Driver type: vmdk {{(pid=61906) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1124.146223] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbad5f80-74b1-4de3-a428-53aa48cdd965 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284841', 'volume_id': 'ae96854f-1414-4b09-bb3e-e3014875951a', 'name': 'volume-ae96854f-1414-4b09-bb3e-e3014875951a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '256457e7-8d94-46f8-8edb-c5782ac34225', 'attached_at': '', 'detached_at': '', 'volume_id': 'ae96854f-1414-4b09-bb3e-e3014875951a', 'serial': 'ae96854f-1414-4b09-bb3e-e3014875951a'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1124.147564] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bc009b5-ff01-4a9a-a306-abdaf5f5f8a8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.182630] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-676b475b-5984-45aa-96f1-bfaf349d1ff0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.190499] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29499515-5592-4437-8551-6cbd0107fb12 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.215843] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa302d93-c853-45e1-87fe-6cf9cdfb7608 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.231600] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbad5f80-74b1-4de3-a428-53aa48cdd965 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] The volume has not been displaced from its original location: [datastore1] volume-ae96854f-1414-4b09-bb3e-e3014875951a/volume-ae96854f-1414-4b09-bb3e-e3014875951a.vmdk. No consolidation needed. {{(pid=61906) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1124.237033] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbad5f80-74b1-4de3-a428-53aa48cdd965 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Reconfiguring VM instance instance-00000066 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1124.237456] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff5037d0-c7fe-45e9-9a19-487a7328dcec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.256222] env[61906]: DEBUG oslo_vmware.api [None req-dbad5f80-74b1-4de3-a428-53aa48cdd965 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 1124.256222] env[61906]: value = "task-1333755" [ 1124.256222] env[61906]: _type = "Task" [ 1124.256222] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.264850] env[61906]: DEBUG oslo_vmware.api [None req-dbad5f80-74b1-4de3-a428-53aa48cdd965 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333755, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.368761] env[61906]: DEBUG oslo_vmware.api [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333754, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.529679] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 9f1253e6-4727-4dca-bad2-04f0c5424cdb] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1124.766910] env[61906]: DEBUG oslo_vmware.api [None req-dbad5f80-74b1-4de3-a428-53aa48cdd965 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333755, 'name': ReconfigVM_Task, 'duration_secs': 0.346373} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.767241] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbad5f80-74b1-4de3-a428-53aa48cdd965 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Reconfigured VM instance instance-00000066 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1124.771865] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-96837f32-3c2b-44c6-ac35-f7079468e5b8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.786988] env[61906]: DEBUG oslo_vmware.api [None req-dbad5f80-74b1-4de3-a428-53aa48cdd965 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 1124.786988] env[61906]: value = "task-1333756" [ 1124.786988] env[61906]: _type = "Task" [ 1124.786988] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.794778] env[61906]: DEBUG oslo_vmware.api [None req-dbad5f80-74b1-4de3-a428-53aa48cdd965 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333756, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.869193] env[61906]: DEBUG oslo_vmware.api [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333754, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.554533} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.869474] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] e75681a9-c315-4f60-9095-fa60e7e39c22/e75681a9-c315-4f60-9095-fa60e7e39c22.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1124.869700] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1124.869964] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7f8b50d1-907f-48fc-8ef9-e182821e28b8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.876634] env[61906]: DEBUG oslo_vmware.api [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1124.876634] env[61906]: value = "task-1333757" [ 1124.876634] env[61906]: _type = "Task" [ 1124.876634] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.884244] env[61906]: DEBUG oslo_vmware.api [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333757, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.035882] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 7e05fee8-842f-4e03-853a-58bac5a9ac0e] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1125.297497] env[61906]: DEBUG oslo_vmware.api [None req-dbad5f80-74b1-4de3-a428-53aa48cdd965 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333756, 'name': ReconfigVM_Task, 'duration_secs': 0.172438} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.297806] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbad5f80-74b1-4de3-a428-53aa48cdd965 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284841', 'volume_id': 'ae96854f-1414-4b09-bb3e-e3014875951a', 'name': 'volume-ae96854f-1414-4b09-bb3e-e3014875951a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '256457e7-8d94-46f8-8edb-c5782ac34225', 'attached_at': '', 'detached_at': '', 'volume_id': 'ae96854f-1414-4b09-bb3e-e3014875951a', 'serial': 'ae96854f-1414-4b09-bb3e-e3014875951a'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1125.388688] env[61906]: DEBUG oslo_vmware.api [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333757, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065537} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.388979] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1125.389797] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04eed95f-6a61-481d-b5b7-973dcaa1cfed {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.411567] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] e75681a9-c315-4f60-9095-fa60e7e39c22/e75681a9-c315-4f60-9095-fa60e7e39c22.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1125.411860] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1e48aa30-1724-474f-b818-3eca2f8e4fe2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.431643] env[61906]: DEBUG oslo_vmware.api [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1125.431643] env[61906]: value = "task-1333758" [ 1125.431643] env[61906]: _type = "Task" [ 1125.431643] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.438959] env[61906]: DEBUG oslo_vmware.api [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333758, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.543822] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 7e452659-0a5e-4740-b7ed-1fd7b1917dac] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1125.838869] env[61906]: DEBUG nova.objects.instance [None req-dbad5f80-74b1-4de3-a428-53aa48cdd965 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lazy-loading 'flavor' on Instance uuid 256457e7-8d94-46f8-8edb-c5782ac34225 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1125.941950] env[61906]: DEBUG oslo_vmware.api [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333758, 'name': ReconfigVM_Task, 'duration_secs': 0.267058} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.942295] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Reconfigured VM instance instance-0000006b to attach disk [datastore2] e75681a9-c315-4f60-9095-fa60e7e39c22/e75681a9-c315-4f60-9095-fa60e7e39c22.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1125.942979] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3ca15031-e430-4147-a0cb-b1643548b54a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.949180] env[61906]: DEBUG oslo_vmware.api [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1125.949180] env[61906]: value = "task-1333759" [ 1125.949180] env[61906]: _type = "Task" [ 1125.949180] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.957009] env[61906]: DEBUG oslo_vmware.api [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333759, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.047454] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1126.047769] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Cleaning up deleted instances with incomplete migration {{(pid=61906) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1126.459328] env[61906]: DEBUG oslo_vmware.api [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333759, 'name': Rename_Task, 'duration_secs': 0.138696} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.459557] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1126.459810] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6bdab8f1-6964-4c6e-a6d4-984875b6f3e5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.465701] env[61906]: DEBUG oslo_vmware.api [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1126.465701] env[61906]: value = "task-1333760" [ 1126.465701] env[61906]: _type = "Task" [ 1126.465701] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.472530] env[61906]: DEBUG oslo_vmware.api [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333760, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.550127] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1126.849174] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ced329dc-1275-46b7-9e6b-e71570fc3681 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "256457e7-8d94-46f8-8edb-c5782ac34225" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.851797] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dbad5f80-74b1-4de3-a428-53aa48cdd965 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "256457e7-8d94-46f8-8edb-c5782ac34225" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.250s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.853052] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ced329dc-1275-46b7-9e6b-e71570fc3681 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "256457e7-8d94-46f8-8edb-c5782ac34225" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.004s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.876551] env[61906]: DEBUG oslo_vmware.rw_handles [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b1a53b-cc8f-07b4-76c8-6c7e8924607f/disk-0.vmdk. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1126.877776] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf2f1b2f-a0dd-4d32-82e7-6bb203102789 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.884802] env[61906]: DEBUG oslo_vmware.rw_handles [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b1a53b-cc8f-07b4-76c8-6c7e8924607f/disk-0.vmdk is in state: ready. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1126.884973] env[61906]: ERROR oslo_vmware.rw_handles [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b1a53b-cc8f-07b4-76c8-6c7e8924607f/disk-0.vmdk due to incomplete transfer. [ 1126.885439] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-16f061a0-146b-4002-84b2-85130440dda7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.893279] env[61906]: DEBUG oslo_vmware.rw_handles [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b1a53b-cc8f-07b4-76c8-6c7e8924607f/disk-0.vmdk. {{(pid=61906) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1126.893482] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Uploaded image f6304fcd-34fc-4c19-a45b-e5229f8ce75a to the Glance image server {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1126.895815] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Destroying the VM {{(pid=61906) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1126.896077] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c9ed0cf9-81f6-4ea3-b1a6-fcea0d41e2a2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.902875] env[61906]: DEBUG oslo_vmware.api [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1126.902875] env[61906]: value = "task-1333761" [ 1126.902875] env[61906]: _type = "Task" [ 1126.902875] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.911741] env[61906]: DEBUG oslo_vmware.api [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333761, 'name': Destroy_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.975284] env[61906]: DEBUG oslo_vmware.api [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333760, 'name': PowerOnVM_Task, 'duration_secs': 0.469054} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.975519] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1126.975721] env[61906]: INFO nova.compute.manager [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Took 7.26 seconds to spawn the instance on the hypervisor. [ 1126.975856] env[61906]: DEBUG nova.compute.manager [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1126.976639] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b414292-f5ce-4aa7-8341-14f75ddb9f67 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.360134] env[61906]: INFO nova.compute.manager [None req-ced329dc-1275-46b7-9e6b-e71570fc3681 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Detaching volume 4d0201ec-4b0a-4f73-a89a-9f1c8a0b6b50 [ 1127.390296] env[61906]: INFO nova.virt.block_device [None req-ced329dc-1275-46b7-9e6b-e71570fc3681 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Attempting to driver detach volume 4d0201ec-4b0a-4f73-a89a-9f1c8a0b6b50 from mountpoint /dev/sdc [ 1127.390672] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-ced329dc-1275-46b7-9e6b-e71570fc3681 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Volume detach. Driver type: vmdk {{(pid=61906) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1127.390739] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-ced329dc-1275-46b7-9e6b-e71570fc3681 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284842', 'volume_id': '4d0201ec-4b0a-4f73-a89a-9f1c8a0b6b50', 'name': 'volume-4d0201ec-4b0a-4f73-a89a-9f1c8a0b6b50', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '256457e7-8d94-46f8-8edb-c5782ac34225', 'attached_at': '', 'detached_at': '', 'volume_id': '4d0201ec-4b0a-4f73-a89a-9f1c8a0b6b50', 'serial': '4d0201ec-4b0a-4f73-a89a-9f1c8a0b6b50'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1127.391713] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc1a87f8-949a-4209-b3d2-23728841bf53 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.416538] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a98147f8-f15d-4fa7-9772-192964bef1e3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.423861] env[61906]: DEBUG oslo_vmware.api [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333761, 'name': Destroy_Task, 'duration_secs': 0.427537} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.425530] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Destroyed the VM [ 1127.425776] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Deleting Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1127.426124] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-245e19c5-3e70-419b-b81e-63aee95605bb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.428056] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48cb1b58-aeb7-441c-a9f0-55fc123ff236 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.450999] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a840d577-0f30-4ddb-beea-11924f032946 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.453736] env[61906]: DEBUG oslo_vmware.api [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1127.453736] env[61906]: value = "task-1333762" [ 1127.453736] env[61906]: _type = "Task" [ 1127.453736] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.461830] env[61906]: DEBUG oslo_vmware.api [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333762, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.471817] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-ced329dc-1275-46b7-9e6b-e71570fc3681 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] The volume has not been displaced from its original location: [datastore1] volume-4d0201ec-4b0a-4f73-a89a-9f1c8a0b6b50/volume-4d0201ec-4b0a-4f73-a89a-9f1c8a0b6b50.vmdk. No consolidation needed. {{(pid=61906) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1127.477096] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-ced329dc-1275-46b7-9e6b-e71570fc3681 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Reconfiguring VM instance instance-00000066 to detach disk 2002 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1127.477400] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f84bf354-85a8-491f-aa7a-b64c88569bc6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.497865] env[61906]: INFO nova.compute.manager [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Took 12.00 seconds to build instance. [ 1127.500011] env[61906]: DEBUG oslo_vmware.api [None req-ced329dc-1275-46b7-9e6b-e71570fc3681 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 1127.500011] env[61906]: value = "task-1333763" [ 1127.500011] env[61906]: _type = "Task" [ 1127.500011] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.508638] env[61906]: DEBUG oslo_vmware.api [None req-ced329dc-1275-46b7-9e6b-e71570fc3681 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333763, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.630646] env[61906]: DEBUG nova.compute.manager [req-bac98baf-3f99-4e57-8ad0-0bb4cdd4ff08 req-e9689c0d-0083-4062-b871-d00f07e2ae71 service nova] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Received event network-changed-1b29cb2d-b70e-4022-b7c7-fd014b216c1c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1127.631088] env[61906]: DEBUG nova.compute.manager [req-bac98baf-3f99-4e57-8ad0-0bb4cdd4ff08 req-e9689c0d-0083-4062-b871-d00f07e2ae71 service nova] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Refreshing instance network info cache due to event network-changed-1b29cb2d-b70e-4022-b7c7-fd014b216c1c. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1127.631088] env[61906]: DEBUG oslo_concurrency.lockutils [req-bac98baf-3f99-4e57-8ad0-0bb4cdd4ff08 req-e9689c0d-0083-4062-b871-d00f07e2ae71 service nova] Acquiring lock "refresh_cache-e75681a9-c315-4f60-9095-fa60e7e39c22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1127.631207] env[61906]: DEBUG oslo_concurrency.lockutils [req-bac98baf-3f99-4e57-8ad0-0bb4cdd4ff08 req-e9689c0d-0083-4062-b871-d00f07e2ae71 service nova] Acquired lock "refresh_cache-e75681a9-c315-4f60-9095-fa60e7e39c22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.631373] env[61906]: DEBUG nova.network.neutron [req-bac98baf-3f99-4e57-8ad0-0bb4cdd4ff08 req-e9689c0d-0083-4062-b871-d00f07e2ae71 service nova] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Refreshing network info cache for port 1b29cb2d-b70e-4022-b7c7-fd014b216c1c {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1127.963442] env[61906]: DEBUG oslo_vmware.api [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333762, 'name': RemoveSnapshot_Task, 'duration_secs': 0.383029} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.963726] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Deleted Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1127.964025] env[61906]: DEBUG nova.compute.manager [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1127.964805] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb039c81-ac18-41c9-8a7f-9b869e3cfbf6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.000970] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ba3e257c-f949-4271-8e1a-d98899b56277 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "e75681a9-c315-4f60-9095-fa60e7e39c22" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.517s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.010454] env[61906]: DEBUG oslo_vmware.api [None req-ced329dc-1275-46b7-9e6b-e71570fc3681 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333763, 'name': ReconfigVM_Task, 'duration_secs': 0.2691} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.010734] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-ced329dc-1275-46b7-9e6b-e71570fc3681 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Reconfigured VM instance instance-00000066 to detach disk 2002 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1128.015491] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-359a6fef-fa90-44f3-b370-845ee840281a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.030019] env[61906]: DEBUG oslo_vmware.api [None req-ced329dc-1275-46b7-9e6b-e71570fc3681 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 1128.030019] env[61906]: value = "task-1333764" [ 1128.030019] env[61906]: _type = "Task" [ 1128.030019] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.037249] env[61906]: DEBUG oslo_vmware.api [None req-ced329dc-1275-46b7-9e6b-e71570fc3681 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333764, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.047019] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1128.353525] env[61906]: DEBUG nova.network.neutron [req-bac98baf-3f99-4e57-8ad0-0bb4cdd4ff08 req-e9689c0d-0083-4062-b871-d00f07e2ae71 service nova] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Updated VIF entry in instance network info cache for port 1b29cb2d-b70e-4022-b7c7-fd014b216c1c. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1128.353925] env[61906]: DEBUG nova.network.neutron [req-bac98baf-3f99-4e57-8ad0-0bb4cdd4ff08 req-e9689c0d-0083-4062-b871-d00f07e2ae71 service nova] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Updating instance_info_cache with network_info: [{"id": "1b29cb2d-b70e-4022-b7c7-fd014b216c1c", "address": "fa:16:3e:7f:a9:4c", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b29cb2d-b7", "ovs_interfaceid": "1b29cb2d-b70e-4022-b7c7-fd014b216c1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1128.477035] env[61906]: INFO nova.compute.manager [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Shelve offloading [ 1128.478737] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1128.478998] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d958adcc-8f03-4233-ac63-948bf07f4cb3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.486282] env[61906]: DEBUG oslo_vmware.api [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1128.486282] env[61906]: value = "task-1333765" [ 1128.486282] env[61906]: _type = "Task" [ 1128.486282] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.496543] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] VM already powered off {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1128.496741] env[61906]: DEBUG nova.compute.manager [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1128.497488] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9983c484-78b3-474c-b0d2-31f3506f0818 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.502761] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "refresh_cache-151cc6f4-96c5-4ac3-97d5-73cddfa7b465" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.502928] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquired lock "refresh_cache-151cc6f4-96c5-4ac3-97d5-73cddfa7b465" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.503112] env[61906]: DEBUG nova.network.neutron [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1128.538430] env[61906]: DEBUG oslo_vmware.api [None req-ced329dc-1275-46b7-9e6b-e71570fc3681 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333764, 'name': ReconfigVM_Task, 'duration_secs': 0.284228} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.538698] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-ced329dc-1275-46b7-9e6b-e71570fc3681 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284842', 'volume_id': '4d0201ec-4b0a-4f73-a89a-9f1c8a0b6b50', 'name': 'volume-4d0201ec-4b0a-4f73-a89a-9f1c8a0b6b50', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '256457e7-8d94-46f8-8edb-c5782ac34225', 'attached_at': '', 'detached_at': '', 'volume_id': '4d0201ec-4b0a-4f73-a89a-9f1c8a0b6b50', 'serial': '4d0201ec-4b0a-4f73-a89a-9f1c8a0b6b50'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1128.551696] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1128.551923] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Starting heal instance info cache {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1128.856940] env[61906]: DEBUG oslo_concurrency.lockutils [req-bac98baf-3f99-4e57-8ad0-0bb4cdd4ff08 req-e9689c0d-0083-4062-b871-d00f07e2ae71 service nova] Releasing lock "refresh_cache-e75681a9-c315-4f60-9095-fa60e7e39c22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1129.080529] env[61906]: DEBUG nova.objects.instance [None req-ced329dc-1275-46b7-9e6b-e71570fc3681 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lazy-loading 'flavor' on Instance uuid 256457e7-8d94-46f8-8edb-c5782ac34225 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1129.199175] env[61906]: DEBUG nova.network.neutron [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Updating instance_info_cache with network_info: [{"id": "1d0e39a6-9bf4-4034-b7e3-0d500971fc09", "address": "fa:16:3e:2b:07:0a", "network": {"id": "0085e7d2-e26f-4b5d-a282-cf6e0f81786a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1504444833-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.153", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00c03ca0de1447da9d47c83f256efdcd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d0e39a6-9b", "ovs_interfaceid": "1d0e39a6-9bf4-4034-b7e3-0d500971fc09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1129.701701] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Releasing lock "refresh_cache-151cc6f4-96c5-4ac3-97d5-73cddfa7b465" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1129.976215] env[61906]: DEBUG nova.compute.manager [req-0aacb309-9116-4041-8668-ec741ed91f50 req-76b51959-2a70-48c5-a754-a0c2ff4780e8 service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Received event network-vif-unplugged-1d0e39a6-9bf4-4034-b7e3-0d500971fc09 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1129.976551] env[61906]: DEBUG oslo_concurrency.lockutils [req-0aacb309-9116-4041-8668-ec741ed91f50 req-76b51959-2a70-48c5-a754-a0c2ff4780e8 service nova] Acquiring lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.976862] env[61906]: DEBUG oslo_concurrency.lockutils [req-0aacb309-9116-4041-8668-ec741ed91f50 req-76b51959-2a70-48c5-a754-a0c2ff4780e8 service nova] Lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.977128] env[61906]: DEBUG oslo_concurrency.lockutils [req-0aacb309-9116-4041-8668-ec741ed91f50 req-76b51959-2a70-48c5-a754-a0c2ff4780e8 service nova] Lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.977388] env[61906]: DEBUG nova.compute.manager [req-0aacb309-9116-4041-8668-ec741ed91f50 req-76b51959-2a70-48c5-a754-a0c2ff4780e8 service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] No waiting events found dispatching network-vif-unplugged-1d0e39a6-9bf4-4034-b7e3-0d500971fc09 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1129.977651] env[61906]: WARNING nova.compute.manager [req-0aacb309-9116-4041-8668-ec741ed91f50 req-76b51959-2a70-48c5-a754-a0c2ff4780e8 service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Received unexpected event network-vif-unplugged-1d0e39a6-9bf4-4034-b7e3-0d500971fc09 for instance with vm_state shelved and task_state shelving_offloading. [ 1130.087753] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ced329dc-1275-46b7-9e6b-e71570fc3681 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "256457e7-8d94-46f8-8edb-c5782ac34225" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.235s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.118842] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1130.119812] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73d10588-e0e5-4a3c-b3dc-c5844e2815ff {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.128224] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1130.128452] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cab812ac-5797-424e-abd2-30d351c24421 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.396186] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1130.396374] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1130.396566] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Deleting the datastore file [datastore1] 151cc6f4-96c5-4ac3-97d5-73cddfa7b465 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1130.396847] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d20e5152-c520-4727-a854-9782ee1822d9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.403923] env[61906]: DEBUG oslo_vmware.api [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1130.403923] env[61906]: value = "task-1333767" [ 1130.403923] env[61906]: _type = "Task" [ 1130.403923] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.412906] env[61906]: DEBUG oslo_vmware.api [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333767, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.593362] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "refresh_cache-86be7197-c0d1-49c5-aa30-cf908a506031" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1130.593531] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquired lock "refresh_cache-86be7197-c0d1-49c5-aa30-cf908a506031" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1130.593714] env[61906]: DEBUG nova.network.neutron [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Forcefully refreshing network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1130.914463] env[61906]: DEBUG oslo_vmware.api [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333767, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.217036} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.914760] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1130.914959] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1130.915160] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1130.942174] env[61906]: INFO nova.scheduler.client.report [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Deleted allocations for instance 151cc6f4-96c5-4ac3-97d5-73cddfa7b465 [ 1131.277432] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "256457e7-8d94-46f8-8edb-c5782ac34225" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.277725] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "256457e7-8d94-46f8-8edb-c5782ac34225" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.277821] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "256457e7-8d94-46f8-8edb-c5782ac34225-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.278023] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "256457e7-8d94-46f8-8edb-c5782ac34225-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.278179] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "256457e7-8d94-46f8-8edb-c5782ac34225-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.280341] env[61906]: INFO nova.compute.manager [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Terminating instance [ 1131.282094] env[61906]: DEBUG nova.compute.manager [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1131.282296] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1131.283161] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9c027c5-534f-4120-bd76-4468fb346058 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.290432] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1131.290681] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5f453990-d32d-49c3-b0c2-02fb762fe408 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.297792] env[61906]: DEBUG oslo_vmware.api [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 1131.297792] env[61906]: value = "task-1333768" [ 1131.297792] env[61906]: _type = "Task" [ 1131.297792] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.305130] env[61906]: DEBUG oslo_vmware.api [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333768, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.446872] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.447205] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.447448] env[61906]: DEBUG nova.objects.instance [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lazy-loading 'resources' on Instance uuid 151cc6f4-96c5-4ac3-97d5-73cddfa7b465 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1131.809675] env[61906]: DEBUG oslo_vmware.api [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333768, 'name': PowerOffVM_Task, 'duration_secs': 0.263233} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.809896] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1131.810082] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1131.810344] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dafeb167-4b1a-4c45-a3cf-da50fa8ca450 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.873730] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1131.873968] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1131.874175] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Deleting the datastore file [datastore1] 256457e7-8d94-46f8-8edb-c5782ac34225 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1131.874443] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-96a95a1b-25e9-42cd-a02f-3870b873c526 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.881414] env[61906]: DEBUG oslo_vmware.api [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for the task: (returnval){ [ 1131.881414] env[61906]: value = "task-1333770" [ 1131.881414] env[61906]: _type = "Task" [ 1131.881414] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.889742] env[61906]: DEBUG oslo_vmware.api [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333770, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.949951] env[61906]: DEBUG nova.objects.instance [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lazy-loading 'numa_topology' on Instance uuid 151cc6f4-96c5-4ac3-97d5-73cddfa7b465 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1132.005398] env[61906]: DEBUG nova.compute.manager [req-e273f03f-022e-41e5-8d39-aa3e40ec25cb req-79822a9d-a9d4-4372-87bf-db8651fe4b35 service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Received event network-changed-1d0e39a6-9bf4-4034-b7e3-0d500971fc09 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1132.005670] env[61906]: DEBUG nova.compute.manager [req-e273f03f-022e-41e5-8d39-aa3e40ec25cb req-79822a9d-a9d4-4372-87bf-db8651fe4b35 service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Refreshing instance network info cache due to event network-changed-1d0e39a6-9bf4-4034-b7e3-0d500971fc09. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1132.005915] env[61906]: DEBUG oslo_concurrency.lockutils [req-e273f03f-022e-41e5-8d39-aa3e40ec25cb req-79822a9d-a9d4-4372-87bf-db8651fe4b35 service nova] Acquiring lock "refresh_cache-151cc6f4-96c5-4ac3-97d5-73cddfa7b465" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1132.006150] env[61906]: DEBUG oslo_concurrency.lockutils [req-e273f03f-022e-41e5-8d39-aa3e40ec25cb req-79822a9d-a9d4-4372-87bf-db8651fe4b35 service nova] Acquired lock "refresh_cache-151cc6f4-96c5-4ac3-97d5-73cddfa7b465" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.006346] env[61906]: DEBUG nova.network.neutron [req-e273f03f-022e-41e5-8d39-aa3e40ec25cb req-79822a9d-a9d4-4372-87bf-db8651fe4b35 service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Refreshing network info cache for port 1d0e39a6-9bf4-4034-b7e3-0d500971fc09 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1132.010776] env[61906]: DEBUG nova.network.neutron [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Updating instance_info_cache with network_info: [{"id": "24ab450e-e9c4-4d9a-9c4d-13dee30eb56b", "address": "fa:16:3e:fe:51:59", "network": {"id": "9903d8f6-78ef-4ec6-8386-c031ceb715f6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1583036471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d6fb2749df6498f82aaf7c5529b309f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24ab450e-e9", "ovs_interfaceid": "24ab450e-e9c4-4d9a-9c4d-13dee30eb56b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1132.141390] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c79f0e28-bb09-4653-93ff-ec6ed44c04bc tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "d4a6abdd-75ee-4040-8ca8-1863dd40695c" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.141646] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c79f0e28-bb09-4653-93ff-ec6ed44c04bc tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "d4a6abdd-75ee-4040-8ca8-1863dd40695c" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.392102] env[61906]: DEBUG oslo_vmware.api [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Task: {'id': task-1333770, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133909} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.392511] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1132.392559] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1132.392775] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1132.392960] env[61906]: INFO nova.compute.manager [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1132.393241] env[61906]: DEBUG oslo.service.loopingcall [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1132.393445] env[61906]: DEBUG nova.compute.manager [-] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1132.393542] env[61906]: DEBUG nova.network.neutron [-] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1132.452831] env[61906]: DEBUG nova.objects.base [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Object Instance<151cc6f4-96c5-4ac3-97d5-73cddfa7b465> lazy-loaded attributes: resources,numa_topology {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1132.513416] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Releasing lock "refresh_cache-86be7197-c0d1-49c5-aa30-cf908a506031" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.513622] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Updated the network info_cache for instance {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1132.514127] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1132.514316] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1132.514494] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1132.514671] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1132.514830] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1132.514967] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61906) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1132.515132] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1132.544856] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3455cc7-bda8-4b62-80d8-59264174330b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.552470] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51c26343-5cbc-4be8-bddc-18ca4921c3e2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.587963] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50e31f96-099a-43f5-baee-5402c09139ee {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.595685] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a954380-ebfe-4759-afe5-ef6d4580e4a7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.608945] env[61906]: DEBUG nova.compute.provider_tree [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1132.644383] env[61906]: DEBUG nova.compute.utils [None req-c79f0e28-bb09-4653-93ff-ec6ed44c04bc tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1132.668489] env[61906]: DEBUG oslo_concurrency.lockutils [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.793441] env[61906]: DEBUG nova.network.neutron [req-e273f03f-022e-41e5-8d39-aa3e40ec25cb req-79822a9d-a9d4-4372-87bf-db8651fe4b35 service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Updated VIF entry in instance network info cache for port 1d0e39a6-9bf4-4034-b7e3-0d500971fc09. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1132.794530] env[61906]: DEBUG nova.network.neutron [req-e273f03f-022e-41e5-8d39-aa3e40ec25cb req-79822a9d-a9d4-4372-87bf-db8651fe4b35 service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Updating instance_info_cache with network_info: [{"id": "1d0e39a6-9bf4-4034-b7e3-0d500971fc09", "address": "fa:16:3e:2b:07:0a", "network": {"id": "0085e7d2-e26f-4b5d-a282-cf6e0f81786a", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1504444833-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.153", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00c03ca0de1447da9d47c83f256efdcd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap1d0e39a6-9b", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1133.018639] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.111873] env[61906]: DEBUG nova.scheduler.client.report [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1133.146790] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c79f0e28-bb09-4653-93ff-ec6ed44c04bc tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "d4a6abdd-75ee-4040-8ca8-1863dd40695c" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.296600] env[61906]: DEBUG oslo_concurrency.lockutils [req-e273f03f-022e-41e5-8d39-aa3e40ec25cb req-79822a9d-a9d4-4372-87bf-db8651fe4b35 service nova] Releasing lock "refresh_cache-151cc6f4-96c5-4ac3-97d5-73cddfa7b465" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1133.299170] env[61906]: DEBUG nova.network.neutron [-] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1133.617019] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.170s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.619512] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.601s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.620205] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.620205] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61906) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1133.620927] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3593443-959e-41a2-8b89-f158104597e0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.632523] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c16bd06-3a6d-4acb-babc-9a50306b01b9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.646930] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1690cc4-ecc1-4626-8fa4-9bb860c0c05b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.653851] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fe47e05-5eea-40e0-b1e8-35be54e83d7e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.684893] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180106MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=61906) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1133.685075] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.685276] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.802396] env[61906]: INFO nova.compute.manager [-] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Took 1.41 seconds to deallocate network for instance. [ 1134.031343] env[61906]: DEBUG nova.compute.manager [req-b7697c6b-b9a6-463b-ab5c-bf8023584043 req-0e4fe260-921d-4ad5-a332-eb3578ce1073 service nova] [instance: 256457e7-8d94-46f8-8edb-c5782ac34225] Received event network-vif-deleted-261ecf70-132a-4fe2-be6b-556a555c86cb {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1134.125637] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a8d7cbf1-ad5e-40ae-9069-4ff876a70cb8 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 20.665s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.126494] env[61906]: DEBUG oslo_concurrency.lockutils [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.458s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.126724] env[61906]: INFO nova.compute.manager [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Unshelving [ 1134.210499] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c79f0e28-bb09-4653-93ff-ec6ed44c04bc tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "d4a6abdd-75ee-4040-8ca8-1863dd40695c" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.210753] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c79f0e28-bb09-4653-93ff-ec6ed44c04bc tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "d4a6abdd-75ee-4040-8ca8-1863dd40695c" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.211009] env[61906]: INFO nova.compute.manager [None req-c79f0e28-bb09-4653-93ff-ec6ed44c04bc tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Attaching volume cd58e0dc-6ca1-45b1-94fd-4b87651291e9 to /dev/sdb [ 1134.242523] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f24c8921-9667-4b56-9040-542978e343e1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.249789] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea9791b1-71a6-4c67-9861-b88f2c93e46e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.263275] env[61906]: DEBUG nova.virt.block_device [None req-c79f0e28-bb09-4653-93ff-ec6ed44c04bc tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Updating existing volume attachment record: f17745e1-6ddd-4f06-a03c-33bd6bbd63fd {{(pid=61906) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1134.309343] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.714260] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 86be7197-c0d1-49c5-aa30-cf908a506031 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1134.714564] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1134.714564] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 256457e7-8d94-46f8-8edb-c5782ac34225 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1134.714722] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance d4a6abdd-75ee-4040-8ca8-1863dd40695c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1134.714853] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 4146c49d-fd4b-49f6-8b17-76d06fd143af actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1134.714976] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance e75681a9-c315-4f60-9095-fa60e7e39c22 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1135.136181] env[61906]: DEBUG nova.compute.utils [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1135.218062] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 151cc6f4-96c5-4ac3-97d5-73cddfa7b465 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1135.218311] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1135.218460] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1135.299786] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e679c9d9-b898-438d-950f-031322c9c2e0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.308690] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51a5458f-5c60-472f-8ec2-01f23fd448a1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.336970] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7072c98-5003-48b5-ab7d-98fd00915790 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.343819] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36e3a15d-e1f7-498b-a139-d4361b5c8896 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.356259] env[61906]: DEBUG nova.compute.provider_tree [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1135.639689] env[61906]: INFO nova.virt.block_device [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Booting with volume 3e22fc47-57b1-4c7b-b334-e0e05317199f at /dev/sdb [ 1135.668891] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cf0df4d4-9e5d-44b2-bbc2-96196d19ab61 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.677635] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b44f2598-03e1-4b12-b469-698d3a6260bd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.702162] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-799cc8b0-b4fd-4fb6-a4db-049eb990132c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.709479] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34e6df16-e7e9-4da9-9425-c8f302688942 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.735301] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6afc2db-0031-41ee-8e2b-598f9bca7d5c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.741336] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0a1e135-527d-4851-b5f4-ed59c9000dec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.753924] env[61906]: DEBUG nova.virt.block_device [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Updating existing volume attachment record: a0519a8c-cc28-4f0f-9cd8-28ee1f3b1a8a {{(pid=61906) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1135.859113] env[61906]: DEBUG nova.scheduler.client.report [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1136.365949] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1136.366230] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.681s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.366498] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.057s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.366733] env[61906]: DEBUG nova.objects.instance [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lazy-loading 'resources' on Instance uuid 256457e7-8d94-46f8-8edb-c5782ac34225 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1136.951940] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d19cf35e-6b12-4be1-a18a-2c9c8f0c76ae {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.959327] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6407272e-45ae-4cd6-94da-19a38507b46f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.988281] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64949943-e076-43dc-af98-163d38b1107b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.995096] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da9bf4ab-a398-4ea1-828f-c607c43c9375 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.008733] env[61906]: DEBUG nova.compute.provider_tree [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1137.512460] env[61906]: DEBUG nova.scheduler.client.report [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1138.017467] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.651s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.044038] env[61906]: INFO nova.scheduler.client.report [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Deleted allocations for instance 256457e7-8d94-46f8-8edb-c5782ac34225 [ 1138.553041] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c83b0908-7c4e-4dca-ae97-342153a5fa05 tempest-AttachVolumeTestJSON-355920053 tempest-AttachVolumeTestJSON-355920053-project-member] Lock "256457e7-8d94-46f8-8edb-c5782ac34225" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.275s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.806739] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-c79f0e28-bb09-4653-93ff-ec6ed44c04bc tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Volume attach. Driver type: vmdk {{(pid=61906) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1138.807019] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-c79f0e28-bb09-4653-93ff-ec6ed44c04bc tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284848', 'volume_id': 'cd58e0dc-6ca1-45b1-94fd-4b87651291e9', 'name': 'volume-cd58e0dc-6ca1-45b1-94fd-4b87651291e9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd4a6abdd-75ee-4040-8ca8-1863dd40695c', 'attached_at': '', 'detached_at': '', 'volume_id': 'cd58e0dc-6ca1-45b1-94fd-4b87651291e9', 'serial': 'cd58e0dc-6ca1-45b1-94fd-4b87651291e9'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1138.807915] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-523fe742-5af0-4567-8fc2-964d0c4dcd4b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.823622] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a569328f-562c-4f62-955f-693a2e390be2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.847765] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-c79f0e28-bb09-4653-93ff-ec6ed44c04bc tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] volume-cd58e0dc-6ca1-45b1-94fd-4b87651291e9/volume-cd58e0dc-6ca1-45b1-94fd-4b87651291e9.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1138.848074] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fbe7f4db-5463-45dd-9a25-0e7fcf66c7a3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.866016] env[61906]: DEBUG oslo_vmware.api [None req-c79f0e28-bb09-4653-93ff-ec6ed44c04bc tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1138.866016] env[61906]: value = "task-1333779" [ 1138.866016] env[61906]: _type = "Task" [ 1138.866016] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.875977] env[61906]: DEBUG oslo_vmware.api [None req-c79f0e28-bb09-4653-93ff-ec6ed44c04bc tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333779, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.379447] env[61906]: DEBUG oslo_vmware.api [None req-c79f0e28-bb09-4653-93ff-ec6ed44c04bc tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333779, 'name': ReconfigVM_Task, 'duration_secs': 0.326526} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.379861] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-c79f0e28-bb09-4653-93ff-ec6ed44c04bc tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Reconfigured VM instance instance-00000069 to attach disk [datastore1] volume-cd58e0dc-6ca1-45b1-94fd-4b87651291e9/volume-cd58e0dc-6ca1-45b1-94fd-4b87651291e9.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1139.387522] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e80cb007-23e6-42af-acae-f86e79b91ac1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.409500] env[61906]: DEBUG oslo_vmware.api [None req-c79f0e28-bb09-4653-93ff-ec6ed44c04bc tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1139.409500] env[61906]: value = "task-1333781" [ 1139.409500] env[61906]: _type = "Task" [ 1139.409500] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.420237] env[61906]: DEBUG oslo_vmware.api [None req-c79f0e28-bb09-4653-93ff-ec6ed44c04bc tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333781, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.919050] env[61906]: DEBUG oslo_vmware.api [None req-c79f0e28-bb09-4653-93ff-ec6ed44c04bc tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333781, 'name': ReconfigVM_Task, 'duration_secs': 0.136924} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.919050] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-c79f0e28-bb09-4653-93ff-ec6ed44c04bc tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284848', 'volume_id': 'cd58e0dc-6ca1-45b1-94fd-4b87651291e9', 'name': 'volume-cd58e0dc-6ca1-45b1-94fd-4b87651291e9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd4a6abdd-75ee-4040-8ca8-1863dd40695c', 'attached_at': '', 'detached_at': '', 'volume_id': 'cd58e0dc-6ca1-45b1-94fd-4b87651291e9', 'serial': 'cd58e0dc-6ca1-45b1-94fd-4b87651291e9'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1140.953272] env[61906]: DEBUG nova.objects.instance [None req-c79f0e28-bb09-4653-93ff-ec6ed44c04bc tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lazy-loading 'flavor' on Instance uuid d4a6abdd-75ee-4040-8ca8-1863dd40695c {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1141.346359] env[61906]: DEBUG oslo_concurrency.lockutils [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.346641] env[61906]: DEBUG oslo_concurrency.lockutils [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.346860] env[61906]: DEBUG nova.objects.instance [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lazy-loading 'pci_requests' on Instance uuid 151cc6f4-96c5-4ac3-97d5-73cddfa7b465 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1141.458374] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c79f0e28-bb09-4653-93ff-ec6ed44c04bc tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "d4a6abdd-75ee-4040-8ca8-1863dd40695c" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.247s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.513535] env[61906]: INFO nova.compute.manager [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Rescuing [ 1141.513790] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "refresh_cache-d4a6abdd-75ee-4040-8ca8-1863dd40695c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.513946] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquired lock "refresh_cache-d4a6abdd-75ee-4040-8ca8-1863dd40695c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.514133] env[61906]: DEBUG nova.network.neutron [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1141.850737] env[61906]: DEBUG nova.objects.instance [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lazy-loading 'numa_topology' on Instance uuid 151cc6f4-96c5-4ac3-97d5-73cddfa7b465 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1142.265864] env[61906]: DEBUG nova.network.neutron [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Updating instance_info_cache with network_info: [{"id": "8c2c45ef-6d72-4a38-bce1-9cbc751f3274", "address": "fa:16:3e:5c:a3:d4", "network": {"id": "9903d8f6-78ef-4ec6-8386-c031ceb715f6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1583036471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d6fb2749df6498f82aaf7c5529b309f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c2c45ef-6d", "ovs_interfaceid": "8c2c45ef-6d72-4a38-bce1-9cbc751f3274", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1142.353224] env[61906]: INFO nova.compute.claims [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1142.769380] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Releasing lock "refresh_cache-d4a6abdd-75ee-4040-8ca8-1863dd40695c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1143.305098] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1143.305098] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-20db9516-f03f-4ea4-9c48-4f86b0ad8c10 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.312017] env[61906]: DEBUG oslo_vmware.api [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1143.312017] env[61906]: value = "task-1333783" [ 1143.312017] env[61906]: _type = "Task" [ 1143.312017] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.320417] env[61906]: DEBUG oslo_vmware.api [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333783, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.453843] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-496a6c61-6fb0-40f6-8389-7a2c8a96054f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.466506] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e38908c-215e-4cde-b726-a17888e86137 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.499104] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac7e89b4-f497-4fd3-8719-0824b117099c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.506525] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c154f26-a2c9-4e5b-bb81-035e5f535f08 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.519742] env[61906]: DEBUG nova.compute.provider_tree [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1143.821410] env[61906]: DEBUG oslo_vmware.api [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333783, 'name': PowerOffVM_Task, 'duration_secs': 0.182677} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.821789] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1143.822611] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e404472-cae3-47fb-9969-14d26e470005 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.842169] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f30b5c98-db5c-4830-a16c-f7b64f725483 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.867172] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1143.867444] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-30a3440e-c36f-4772-be52-6e98ab42790f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.873400] env[61906]: DEBUG oslo_vmware.api [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1143.873400] env[61906]: value = "task-1333784" [ 1143.873400] env[61906]: _type = "Task" [ 1143.873400] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.882410] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] VM already powered off {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1143.882628] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1143.882892] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1143.883060] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1143.883248] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1143.883474] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-41ec5384-b2f5-43a7-b463-1f062a54cdb3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.890764] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1143.890945] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1143.891622] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d9e1109-bb05-4f45-bd8e-4bc7acada329 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.896161] env[61906]: DEBUG oslo_vmware.api [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1143.896161] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52567ad9-27a6-9b14-a092-ed9c8b76d0c3" [ 1143.896161] env[61906]: _type = "Task" [ 1143.896161] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.903066] env[61906]: DEBUG oslo_vmware.api [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52567ad9-27a6-9b14-a092-ed9c8b76d0c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.023554] env[61906]: DEBUG nova.scheduler.client.report [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1144.406127] env[61906]: DEBUG oslo_vmware.api [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52567ad9-27a6-9b14-a092-ed9c8b76d0c3, 'name': SearchDatastore_Task, 'duration_secs': 0.007823} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.406926] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-736a189d-7b4e-4a48-a85c-aff6c37e2bd5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.411843] env[61906]: DEBUG oslo_vmware.api [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1144.411843] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52594beb-591f-814e-da50-34605f3c6d63" [ 1144.411843] env[61906]: _type = "Task" [ 1144.411843] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.419231] env[61906]: DEBUG oslo_vmware.api [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52594beb-591f-814e-da50-34605f3c6d63, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.528424] env[61906]: DEBUG oslo_concurrency.lockutils [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.182s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.556548] env[61906]: INFO nova.network.neutron [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Updating port 1d0e39a6-9bf4-4034-b7e3-0d500971fc09 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1144.921768] env[61906]: DEBUG oslo_vmware.api [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52594beb-591f-814e-da50-34605f3c6d63, 'name': SearchDatastore_Task, 'duration_secs': 0.009007} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.922053] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1144.922330] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] d4a6abdd-75ee-4040-8ca8-1863dd40695c/a9502563-1f96-46a5-b87e-12c0e9375e6e-rescue.vmdk. {{(pid=61906) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1144.922599] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e5ef0e91-4b66-4a71-9816-6e75534463f6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.928818] env[61906]: DEBUG oslo_vmware.api [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1144.928818] env[61906]: value = "task-1333785" [ 1144.928818] env[61906]: _type = "Task" [ 1144.928818] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.936150] env[61906]: DEBUG oslo_vmware.api [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333785, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.439135] env[61906]: DEBUG oslo_vmware.api [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333785, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.441133} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.439498] env[61906]: INFO nova.virt.vmwareapi.ds_util [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] d4a6abdd-75ee-4040-8ca8-1863dd40695c/a9502563-1f96-46a5-b87e-12c0e9375e6e-rescue.vmdk. [ 1145.440188] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c12069-3c80-4986-90ca-0722c46d3c4d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.467055] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] d4a6abdd-75ee-4040-8ca8-1863dd40695c/a9502563-1f96-46a5-b87e-12c0e9375e6e-rescue.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1145.467055] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0666742e-e8d8-49c1-aced-246d7550580c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.485197] env[61906]: DEBUG oslo_vmware.api [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1145.485197] env[61906]: value = "task-1333787" [ 1145.485197] env[61906]: _type = "Task" [ 1145.485197] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.492757] env[61906]: DEBUG oslo_vmware.api [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333787, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.946452] env[61906]: DEBUG nova.compute.manager [req-c409804b-01b5-406f-b7eb-4de09e03d1b9 req-2c9b7010-dc08-4244-9365-030fa682b201 service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Received event network-vif-plugged-1d0e39a6-9bf4-4034-b7e3-0d500971fc09 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1145.946681] env[61906]: DEBUG oslo_concurrency.lockutils [req-c409804b-01b5-406f-b7eb-4de09e03d1b9 req-2c9b7010-dc08-4244-9365-030fa682b201 service nova] Acquiring lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.946895] env[61906]: DEBUG oslo_concurrency.lockutils [req-c409804b-01b5-406f-b7eb-4de09e03d1b9 req-2c9b7010-dc08-4244-9365-030fa682b201 service nova] Lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1145.947096] env[61906]: DEBUG oslo_concurrency.lockutils [req-c409804b-01b5-406f-b7eb-4de09e03d1b9 req-2c9b7010-dc08-4244-9365-030fa682b201 service nova] Lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.947296] env[61906]: DEBUG nova.compute.manager [req-c409804b-01b5-406f-b7eb-4de09e03d1b9 req-2c9b7010-dc08-4244-9365-030fa682b201 service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] No waiting events found dispatching network-vif-plugged-1d0e39a6-9bf4-4034-b7e3-0d500971fc09 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1145.947523] env[61906]: WARNING nova.compute.manager [req-c409804b-01b5-406f-b7eb-4de09e03d1b9 req-2c9b7010-dc08-4244-9365-030fa682b201 service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Received unexpected event network-vif-plugged-1d0e39a6-9bf4-4034-b7e3-0d500971fc09 for instance with vm_state shelved_offloaded and task_state spawning. [ 1145.993999] env[61906]: DEBUG oslo_vmware.api [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333787, 'name': ReconfigVM_Task, 'duration_secs': 0.321145} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.994671] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Reconfigured VM instance instance-00000069 to attach disk [datastore1] d4a6abdd-75ee-4040-8ca8-1863dd40695c/a9502563-1f96-46a5-b87e-12c0e9375e6e-rescue.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1145.995610] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bf0924a-4c1e-4757-ac47-11c215fb3b07 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.023462] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1bff729d-d296-44ae-8e38-f4c08fb866ba {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.034619] env[61906]: DEBUG oslo_concurrency.lockutils [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "refresh_cache-151cc6f4-96c5-4ac3-97d5-73cddfa7b465" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1146.034818] env[61906]: DEBUG oslo_concurrency.lockutils [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquired lock "refresh_cache-151cc6f4-96c5-4ac3-97d5-73cddfa7b465" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.035009] env[61906]: DEBUG nova.network.neutron [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1146.038412] env[61906]: DEBUG oslo_vmware.api [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1146.038412] env[61906]: value = "task-1333788" [ 1146.038412] env[61906]: _type = "Task" [ 1146.038412] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.046173] env[61906]: DEBUG oslo_vmware.api [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333788, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.548275] env[61906]: DEBUG oslo_vmware.api [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333788, 'name': ReconfigVM_Task, 'duration_secs': 0.160744} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.548804] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1146.548882] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1df43602-5190-48c2-aba7-fc551270ef58 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.554999] env[61906]: DEBUG oslo_vmware.api [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1146.554999] env[61906]: value = "task-1333789" [ 1146.554999] env[61906]: _type = "Task" [ 1146.554999] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.562410] env[61906]: DEBUG oslo_vmware.api [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333789, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.749646] env[61906]: DEBUG nova.network.neutron [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Updating instance_info_cache with network_info: [{"id": "1d0e39a6-9bf4-4034-b7e3-0d500971fc09", "address": "fa:16:3e:2b:07:0a", "network": {"id": "0085e7d2-e26f-4b5d-a282-cf6e0f81786a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1504444833-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.153", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00c03ca0de1447da9d47c83f256efdcd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d0e39a6-9b", "ovs_interfaceid": "1d0e39a6-9bf4-4034-b7e3-0d500971fc09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1147.064740] env[61906]: DEBUG oslo_vmware.api [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333789, 'name': PowerOnVM_Task, 'duration_secs': 0.379195} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.065072] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1147.068148] env[61906]: DEBUG nova.compute.manager [None req-cd6fb776-d561-4544-b260-b1737b629102 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1147.068920] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ce5a036-ed26-4858-abc8-bea91f5f70ec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.252061] env[61906]: DEBUG oslo_concurrency.lockutils [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Releasing lock "refresh_cache-151cc6f4-96c5-4ac3-97d5-73cddfa7b465" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1147.276663] env[61906]: DEBUG nova.virt.hardware [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='b91a916ee78fe99aa96bd0100d2d5b77',container_format='bare',created_at=2024-10-22T06:29:25Z,direct_url=,disk_format='vmdk',id=f6304fcd-34fc-4c19-a45b-e5229f8ce75a,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1725835554-shelved',owner='00c03ca0de1447da9d47c83f256efdcd',properties=ImageMetaProps,protected=,size=31668736,status='active',tags=,updated_at=2024-10-22T06:29:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1147.276917] env[61906]: DEBUG nova.virt.hardware [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1147.277111] env[61906]: DEBUG nova.virt.hardware [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1147.277303] env[61906]: DEBUG nova.virt.hardware [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1147.277457] env[61906]: DEBUG nova.virt.hardware [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1147.277609] env[61906]: DEBUG nova.virt.hardware [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1147.277822] env[61906]: DEBUG nova.virt.hardware [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1147.277989] env[61906]: DEBUG nova.virt.hardware [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1147.278175] env[61906]: DEBUG nova.virt.hardware [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1147.278343] env[61906]: DEBUG nova.virt.hardware [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1147.278518] env[61906]: DEBUG nova.virt.hardware [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1147.279373] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aff042ea-4ea8-4a27-b34c-7eb1b2645879 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.286863] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e7e7ed1-02c9-4410-af9e-13da09a47382 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.299838] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2b:07:0a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c6934071-bf85-4591-9c7d-55c7ea131262', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1d0e39a6-9bf4-4034-b7e3-0d500971fc09', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1147.307517] env[61906]: DEBUG oslo.service.loopingcall [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1147.307745] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1147.307944] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a303ab7a-2968-457c-8b0d-e35892e6bc66 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.326228] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1147.326228] env[61906]: value = "task-1333790" [ 1147.326228] env[61906]: _type = "Task" [ 1147.326228] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.333066] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333790, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.836060] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333790, 'name': CreateVM_Task, 'duration_secs': 0.331296} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.836424] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1147.836877] env[61906]: DEBUG oslo_concurrency.lockutils [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f6304fcd-34fc-4c19-a45b-e5229f8ce75a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1147.837069] env[61906]: DEBUG oslo_concurrency.lockutils [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f6304fcd-34fc-4c19-a45b-e5229f8ce75a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1147.837466] env[61906]: DEBUG oslo_concurrency.lockutils [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f6304fcd-34fc-4c19-a45b-e5229f8ce75a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1147.837720] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-adf1b0ed-b37a-49d5-aedc-f82c90706f26 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.842239] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1147.842239] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5279f952-ecdb-b8cd-3b36-8731241a05f4" [ 1147.842239] env[61906]: _type = "Task" [ 1147.842239] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.849688] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5279f952-ecdb-b8cd-3b36-8731241a05f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.970784] env[61906]: DEBUG nova.compute.manager [req-63998bf3-899c-4884-93f7-2c80ba9db7be req-508e2497-de17-40f9-9ba2-fbbe6397c10a service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Received event network-changed-1d0e39a6-9bf4-4034-b7e3-0d500971fc09 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1147.970991] env[61906]: DEBUG nova.compute.manager [req-63998bf3-899c-4884-93f7-2c80ba9db7be req-508e2497-de17-40f9-9ba2-fbbe6397c10a service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Refreshing instance network info cache due to event network-changed-1d0e39a6-9bf4-4034-b7e3-0d500971fc09. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1147.971226] env[61906]: DEBUG oslo_concurrency.lockutils [req-63998bf3-899c-4884-93f7-2c80ba9db7be req-508e2497-de17-40f9-9ba2-fbbe6397c10a service nova] Acquiring lock "refresh_cache-151cc6f4-96c5-4ac3-97d5-73cddfa7b465" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1147.971376] env[61906]: DEBUG oslo_concurrency.lockutils [req-63998bf3-899c-4884-93f7-2c80ba9db7be req-508e2497-de17-40f9-9ba2-fbbe6397c10a service nova] Acquired lock "refresh_cache-151cc6f4-96c5-4ac3-97d5-73cddfa7b465" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1147.971538] env[61906]: DEBUG nova.network.neutron [req-63998bf3-899c-4884-93f7-2c80ba9db7be req-508e2497-de17-40f9-9ba2-fbbe6397c10a service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Refreshing network info cache for port 1d0e39a6-9bf4-4034-b7e3-0d500971fc09 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1148.352870] env[61906]: DEBUG oslo_concurrency.lockutils [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f6304fcd-34fc-4c19-a45b-e5229f8ce75a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1148.353210] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Processing image f6304fcd-34fc-4c19-a45b-e5229f8ce75a {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1148.353460] env[61906]: DEBUG oslo_concurrency.lockutils [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f6304fcd-34fc-4c19-a45b-e5229f8ce75a/f6304fcd-34fc-4c19-a45b-e5229f8ce75a.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1148.353621] env[61906]: DEBUG oslo_concurrency.lockutils [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f6304fcd-34fc-4c19-a45b-e5229f8ce75a/f6304fcd-34fc-4c19-a45b-e5229f8ce75a.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1148.353808] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1148.354085] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f1db79eb-3478-4ee7-a6b6-4461321a109b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.370740] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1148.370888] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1148.371646] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ac77e76-c064-4d83-9ac8-475b1b1f2e7d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.376708] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1148.376708] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5282af2f-bbf2-7f2b-ae6a-eecd575226f3" [ 1148.376708] env[61906]: _type = "Task" [ 1148.376708] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.384386] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5282af2f-bbf2-7f2b-ae6a-eecd575226f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.527271] env[61906]: INFO nova.compute.manager [None req-1cd76901-7611-40d5-8f25-0460a481eaea tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Unrescuing [ 1148.527551] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1cd76901-7611-40d5-8f25-0460a481eaea tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "refresh_cache-d4a6abdd-75ee-4040-8ca8-1863dd40695c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1148.527708] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1cd76901-7611-40d5-8f25-0460a481eaea tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquired lock "refresh_cache-d4a6abdd-75ee-4040-8ca8-1863dd40695c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1148.527907] env[61906]: DEBUG nova.network.neutron [None req-1cd76901-7611-40d5-8f25-0460a481eaea tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1148.690011] env[61906]: DEBUG nova.network.neutron [req-63998bf3-899c-4884-93f7-2c80ba9db7be req-508e2497-de17-40f9-9ba2-fbbe6397c10a service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Updated VIF entry in instance network info cache for port 1d0e39a6-9bf4-4034-b7e3-0d500971fc09. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1148.690438] env[61906]: DEBUG nova.network.neutron [req-63998bf3-899c-4884-93f7-2c80ba9db7be req-508e2497-de17-40f9-9ba2-fbbe6397c10a service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Updating instance_info_cache with network_info: [{"id": "1d0e39a6-9bf4-4034-b7e3-0d500971fc09", "address": "fa:16:3e:2b:07:0a", "network": {"id": "0085e7d2-e26f-4b5d-a282-cf6e0f81786a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1504444833-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.153", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00c03ca0de1447da9d47c83f256efdcd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d0e39a6-9b", "ovs_interfaceid": "1d0e39a6-9bf4-4034-b7e3-0d500971fc09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1148.887027] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Preparing fetch location {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1148.887343] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Fetch image to [datastore1] OSTACK_IMG_6fdad45e-284a-420e-83dd-eacbea7fc921/OSTACK_IMG_6fdad45e-284a-420e-83dd-eacbea7fc921.vmdk {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1148.887343] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Downloading stream optimized image f6304fcd-34fc-4c19-a45b-e5229f8ce75a to [datastore1] OSTACK_IMG_6fdad45e-284a-420e-83dd-eacbea7fc921/OSTACK_IMG_6fdad45e-284a-420e-83dd-eacbea7fc921.vmdk on the data store datastore1 as vApp {{(pid=61906) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1148.887440] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Downloading image file data f6304fcd-34fc-4c19-a45b-e5229f8ce75a to the ESX as VM named 'OSTACK_IMG_6fdad45e-284a-420e-83dd-eacbea7fc921' {{(pid=61906) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1148.960899] env[61906]: DEBUG oslo_vmware.rw_handles [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1148.960899] env[61906]: value = "resgroup-9" [ 1148.960899] env[61906]: _type = "ResourcePool" [ 1148.960899] env[61906]: }. {{(pid=61906) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1148.961380] env[61906]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-b67f5bb7-f3f0-42ec-98a6-156297ed5afc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.982600] env[61906]: DEBUG oslo_vmware.rw_handles [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lease: (returnval){ [ 1148.982600] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]524aa2ed-7ae1-3e66-39cd-1b3083a9582f" [ 1148.982600] env[61906]: _type = "HttpNfcLease" [ 1148.982600] env[61906]: } obtained for vApp import into resource pool (val){ [ 1148.982600] env[61906]: value = "resgroup-9" [ 1148.982600] env[61906]: _type = "ResourcePool" [ 1148.982600] env[61906]: }. {{(pid=61906) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1148.982978] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the lease: (returnval){ [ 1148.982978] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]524aa2ed-7ae1-3e66-39cd-1b3083a9582f" [ 1148.982978] env[61906]: _type = "HttpNfcLease" [ 1148.982978] env[61906]: } to be ready. {{(pid=61906) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1148.989855] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1148.989855] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]524aa2ed-7ae1-3e66-39cd-1b3083a9582f" [ 1148.989855] env[61906]: _type = "HttpNfcLease" [ 1148.989855] env[61906]: } is initializing. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1149.194399] env[61906]: DEBUG oslo_concurrency.lockutils [req-63998bf3-899c-4884-93f7-2c80ba9db7be req-508e2497-de17-40f9-9ba2-fbbe6397c10a service nova] Releasing lock "refresh_cache-151cc6f4-96c5-4ac3-97d5-73cddfa7b465" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1149.430272] env[61906]: DEBUG nova.network.neutron [None req-1cd76901-7611-40d5-8f25-0460a481eaea tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Updating instance_info_cache with network_info: [{"id": "8c2c45ef-6d72-4a38-bce1-9cbc751f3274", "address": "fa:16:3e:5c:a3:d4", "network": {"id": "9903d8f6-78ef-4ec6-8386-c031ceb715f6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1583036471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d6fb2749df6498f82aaf7c5529b309f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c2c45ef-6d", "ovs_interfaceid": "8c2c45ef-6d72-4a38-bce1-9cbc751f3274", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1149.493086] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1149.493086] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]524aa2ed-7ae1-3e66-39cd-1b3083a9582f" [ 1149.493086] env[61906]: _type = "HttpNfcLease" [ 1149.493086] env[61906]: } is ready. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1149.493086] env[61906]: DEBUG oslo_vmware.rw_handles [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1149.493086] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]524aa2ed-7ae1-3e66-39cd-1b3083a9582f" [ 1149.493086] env[61906]: _type = "HttpNfcLease" [ 1149.493086] env[61906]: }. {{(pid=61906) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1149.493765] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c37e703e-4ea1-48ed-a796-de61960ebf93 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.501962] env[61906]: DEBUG oslo_vmware.rw_handles [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52358078-2317-3b9d-ea16-f551c82eb69e/disk-0.vmdk from lease info. {{(pid=61906) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1149.502091] env[61906]: DEBUG oslo_vmware.rw_handles [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Creating HTTP connection to write to file with size = 31668736 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52358078-2317-3b9d-ea16-f551c82eb69e/disk-0.vmdk. {{(pid=61906) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1149.568690] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-7c167d9a-011c-4c11-afa2-471206de3e90 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.933589] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1cd76901-7611-40d5-8f25-0460a481eaea tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Releasing lock "refresh_cache-d4a6abdd-75ee-4040-8ca8-1863dd40695c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1149.934393] env[61906]: DEBUG nova.objects.instance [None req-1cd76901-7611-40d5-8f25-0460a481eaea tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lazy-loading 'flavor' on Instance uuid d4a6abdd-75ee-4040-8ca8-1863dd40695c {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1150.440682] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1855a3e0-bc6b-4db1-9001-48ba5760c775 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.465963] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cd76901-7611-40d5-8f25-0460a481eaea tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1150.471124] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4640a92e-de85-4140-8f75-fe1babc71fdd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.480045] env[61906]: DEBUG oslo_vmware.api [None req-1cd76901-7611-40d5-8f25-0460a481eaea tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1150.480045] env[61906]: value = "task-1333792" [ 1150.480045] env[61906]: _type = "Task" [ 1150.480045] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.488703] env[61906]: DEBUG oslo_vmware.api [None req-1cd76901-7611-40d5-8f25-0460a481eaea tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333792, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.688158] env[61906]: DEBUG oslo_vmware.rw_handles [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Completed reading data from the image iterator. {{(pid=61906) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1150.688158] env[61906]: DEBUG oslo_vmware.rw_handles [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52358078-2317-3b9d-ea16-f551c82eb69e/disk-0.vmdk. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1150.688801] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d35875c-af65-4c32-b81d-fbd2bb8ad47a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.696956] env[61906]: DEBUG oslo_vmware.rw_handles [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52358078-2317-3b9d-ea16-f551c82eb69e/disk-0.vmdk is in state: ready. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1150.696956] env[61906]: DEBUG oslo_vmware.rw_handles [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52358078-2317-3b9d-ea16-f551c82eb69e/disk-0.vmdk. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1150.696956] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-15301489-beea-49bd-bd51-3b732ff7d2af {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.884288] env[61906]: DEBUG oslo_vmware.rw_handles [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52358078-2317-3b9d-ea16-f551c82eb69e/disk-0.vmdk. {{(pid=61906) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1150.884529] env[61906]: INFO nova.virt.vmwareapi.images [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Downloaded image file data f6304fcd-34fc-4c19-a45b-e5229f8ce75a [ 1150.885484] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07fb2bb6-7f96-4c49-a5f9-618c68786c5a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.900382] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cdd16679-b510-4a0e-b766-ec59af65052f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.922267] env[61906]: INFO nova.virt.vmwareapi.images [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] The imported VM was unregistered [ 1150.924947] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Caching image {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1150.925196] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Creating directory with path [datastore1] devstack-image-cache_base/f6304fcd-34fc-4c19-a45b-e5229f8ce75a {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1150.925476] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c3e0c6b8-d232-4167-9af1-ab0985cb1cc0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.934564] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Created directory with path [datastore1] devstack-image-cache_base/f6304fcd-34fc-4c19-a45b-e5229f8ce75a {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1150.934866] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_6fdad45e-284a-420e-83dd-eacbea7fc921/OSTACK_IMG_6fdad45e-284a-420e-83dd-eacbea7fc921.vmdk to [datastore1] devstack-image-cache_base/f6304fcd-34fc-4c19-a45b-e5229f8ce75a/f6304fcd-34fc-4c19-a45b-e5229f8ce75a.vmdk. {{(pid=61906) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1150.934970] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-cb1c2cbd-b679-4b73-b2ea-2743db9bba47 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.940800] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1150.940800] env[61906]: value = "task-1333794" [ 1150.940800] env[61906]: _type = "Task" [ 1150.940800] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.948741] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333794, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.991046] env[61906]: DEBUG oslo_vmware.api [None req-1cd76901-7611-40d5-8f25-0460a481eaea tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333792, 'name': PowerOffVM_Task, 'duration_secs': 0.268257} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.991218] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cd76901-7611-40d5-8f25-0460a481eaea tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1150.996661] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-1cd76901-7611-40d5-8f25-0460a481eaea tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Reconfiguring VM instance instance-00000069 to detach disk 2002 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1150.996938] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e95d5e9-4b6d-42d5-97ff-249c701e23f3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.015111] env[61906]: DEBUG oslo_vmware.api [None req-1cd76901-7611-40d5-8f25-0460a481eaea tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1151.015111] env[61906]: value = "task-1333795" [ 1151.015111] env[61906]: _type = "Task" [ 1151.015111] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.027086] env[61906]: DEBUG oslo_vmware.api [None req-1cd76901-7611-40d5-8f25-0460a481eaea tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333795, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.451547] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333794, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.525326] env[61906]: DEBUG oslo_vmware.api [None req-1cd76901-7611-40d5-8f25-0460a481eaea tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333795, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.951937] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333794, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.027772] env[61906]: DEBUG oslo_vmware.api [None req-1cd76901-7611-40d5-8f25-0460a481eaea tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333795, 'name': ReconfigVM_Task, 'duration_secs': 0.704606} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.028131] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-1cd76901-7611-40d5-8f25-0460a481eaea tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Reconfigured VM instance instance-00000069 to detach disk 2002 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1152.028358] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cd76901-7611-40d5-8f25-0460a481eaea tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1152.028635] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6748ed3a-e4a3-4965-9ee4-6f104cb728bd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.034909] env[61906]: DEBUG oslo_vmware.api [None req-1cd76901-7611-40d5-8f25-0460a481eaea tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1152.034909] env[61906]: value = "task-1333796" [ 1152.034909] env[61906]: _type = "Task" [ 1152.034909] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.043065] env[61906]: DEBUG oslo_vmware.api [None req-1cd76901-7611-40d5-8f25-0460a481eaea tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333796, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.452515] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333794, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.545997] env[61906]: DEBUG oslo_vmware.api [None req-1cd76901-7611-40d5-8f25-0460a481eaea tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333796, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.954238] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333794, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.046054] env[61906]: DEBUG oslo_vmware.api [None req-1cd76901-7611-40d5-8f25-0460a481eaea tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333796, 'name': PowerOnVM_Task, 'duration_secs': 0.789078} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.046400] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cd76901-7611-40d5-8f25-0460a481eaea tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1153.046647] env[61906]: DEBUG nova.compute.manager [None req-1cd76901-7611-40d5-8f25-0460a481eaea tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1153.047494] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dabba821-740c-4339-9896-36059b19646b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.454697] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333794, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.955242] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333794, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.992849} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.955622] env[61906]: INFO nova.virt.vmwareapi.ds_util [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_6fdad45e-284a-420e-83dd-eacbea7fc921/OSTACK_IMG_6fdad45e-284a-420e-83dd-eacbea7fc921.vmdk to [datastore1] devstack-image-cache_base/f6304fcd-34fc-4c19-a45b-e5229f8ce75a/f6304fcd-34fc-4c19-a45b-e5229f8ce75a.vmdk. [ 1153.955759] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Cleaning up location [datastore1] OSTACK_IMG_6fdad45e-284a-420e-83dd-eacbea7fc921 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1153.955908] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_6fdad45e-284a-420e-83dd-eacbea7fc921 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1153.956180] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-33405a6d-fd55-420c-a467-9ebda9d3d4ba {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.961694] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1153.961694] env[61906]: value = "task-1333797" [ 1153.961694] env[61906]: _type = "Task" [ 1153.961694] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.968841] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333797, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.214132] env[61906]: DEBUG nova.compute.manager [req-de777ebf-5e21-45a1-aba4-e29eec6d0c20 req-0049fece-4fb3-4032-b7e3-b3008802e6e4 service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Received event network-changed-8c2c45ef-6d72-4a38-bce1-9cbc751f3274 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1154.214416] env[61906]: DEBUG nova.compute.manager [req-de777ebf-5e21-45a1-aba4-e29eec6d0c20 req-0049fece-4fb3-4032-b7e3-b3008802e6e4 service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Refreshing instance network info cache due to event network-changed-8c2c45ef-6d72-4a38-bce1-9cbc751f3274. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1154.214544] env[61906]: DEBUG oslo_concurrency.lockutils [req-de777ebf-5e21-45a1-aba4-e29eec6d0c20 req-0049fece-4fb3-4032-b7e3-b3008802e6e4 service nova] Acquiring lock "refresh_cache-d4a6abdd-75ee-4040-8ca8-1863dd40695c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1154.214695] env[61906]: DEBUG oslo_concurrency.lockutils [req-de777ebf-5e21-45a1-aba4-e29eec6d0c20 req-0049fece-4fb3-4032-b7e3-b3008802e6e4 service nova] Acquired lock "refresh_cache-d4a6abdd-75ee-4040-8ca8-1863dd40695c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1154.214961] env[61906]: DEBUG nova.network.neutron [req-de777ebf-5e21-45a1-aba4-e29eec6d0c20 req-0049fece-4fb3-4032-b7e3-b3008802e6e4 service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Refreshing network info cache for port 8c2c45ef-6d72-4a38-bce1-9cbc751f3274 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1154.471995] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333797, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.267275} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.473138] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1154.473138] env[61906]: DEBUG oslo_concurrency.lockutils [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f6304fcd-34fc-4c19-a45b-e5229f8ce75a/f6304fcd-34fc-4c19-a45b-e5229f8ce75a.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1154.473138] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f6304fcd-34fc-4c19-a45b-e5229f8ce75a/f6304fcd-34fc-4c19-a45b-e5229f8ce75a.vmdk to [datastore1] 151cc6f4-96c5-4ac3-97d5-73cddfa7b465/151cc6f4-96c5-4ac3-97d5-73cddfa7b465.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1154.473293] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-47d11493-bfb5-48f5-99d4-e23796996f17 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.479991] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1154.479991] env[61906]: value = "task-1333798" [ 1154.479991] env[61906]: _type = "Task" [ 1154.479991] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.488311] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333798, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.804663] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._sync_power_states {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1154.995177] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333798, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.107547] env[61906]: DEBUG nova.network.neutron [req-de777ebf-5e21-45a1-aba4-e29eec6d0c20 req-0049fece-4fb3-4032-b7e3-b3008802e6e4 service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Updated VIF entry in instance network info cache for port 8c2c45ef-6d72-4a38-bce1-9cbc751f3274. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1155.108069] env[61906]: DEBUG nova.network.neutron [req-de777ebf-5e21-45a1-aba4-e29eec6d0c20 req-0049fece-4fb3-4032-b7e3-b3008802e6e4 service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Updating instance_info_cache with network_info: [{"id": "8c2c45ef-6d72-4a38-bce1-9cbc751f3274", "address": "fa:16:3e:5c:a3:d4", "network": {"id": "9903d8f6-78ef-4ec6-8386-c031ceb715f6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1583036471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d6fb2749df6498f82aaf7c5529b309f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c2c45ef-6d", "ovs_interfaceid": "8c2c45ef-6d72-4a38-bce1-9cbc751f3274", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1155.309544] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Getting list of instances from cluster (obj){ [ 1155.309544] env[61906]: value = "domain-c8" [ 1155.309544] env[61906]: _type = "ClusterComputeResource" [ 1155.309544] env[61906]: } {{(pid=61906) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1155.311448] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47304668-f36a-43e0-8bc5-875702a83a32 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.328374] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Got total of 6 instances {{(pid=61906) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1155.328658] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Triggering sync for uuid 86be7197-c0d1-49c5-aa30-cf908a506031 {{(pid=61906) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1155.328758] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Triggering sync for uuid f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b {{(pid=61906) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1155.328905] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Triggering sync for uuid 151cc6f4-96c5-4ac3-97d5-73cddfa7b465 {{(pid=61906) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1155.329154] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Triggering sync for uuid d4a6abdd-75ee-4040-8ca8-1863dd40695c {{(pid=61906) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1155.329339] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Triggering sync for uuid 4146c49d-fd4b-49f6-8b17-76d06fd143af {{(pid=61906) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1155.329501] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Triggering sync for uuid e75681a9-c315-4f60-9095-fa60e7e39c22 {{(pid=61906) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1155.329880] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "86be7197-c0d1-49c5-aa30-cf908a506031" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.330216] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "86be7197-c0d1-49c5-aa30-cf908a506031" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.330540] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.330745] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.331100] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.331340] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "d4a6abdd-75ee-4040-8ca8-1863dd40695c" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.331527] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "d4a6abdd-75ee-4040-8ca8-1863dd40695c" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.331854] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "4146c49d-fd4b-49f6-8b17-76d06fd143af" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.332079] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "4146c49d-fd4b-49f6-8b17-76d06fd143af" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.332334] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "e75681a9-c315-4f60-9095-fa60e7e39c22" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.332591] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "e75681a9-c315-4f60-9095-fa60e7e39c22" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.333581] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fd0ba73-f9c1-4c17-b7d5-50d1ee5426ff {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.337254] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9d0e119-9115-4528-829f-46b84accfc98 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.340441] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91066430-c8fc-41d6-91f8-f9e010f98a24 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.343643] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad9bb65f-cd94-4d79-9946-a0bbda035cde {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.346866] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23a6626c-db37-46cf-bab2-c984aa186bfe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.491692] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333798, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.612560] env[61906]: DEBUG oslo_concurrency.lockutils [req-de777ebf-5e21-45a1-aba4-e29eec6d0c20 req-0049fece-4fb3-4032-b7e3-b3008802e6e4 service nova] Releasing lock "refresh_cache-d4a6abdd-75ee-4040-8ca8-1863dd40695c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1155.868919] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "86be7197-c0d1-49c5-aa30-cf908a506031" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.538s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.868919] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "e75681a9-c315-4f60-9095-fa60e7e39c22" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.536s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.868919] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.538s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.874862] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "d4a6abdd-75ee-4040-8ca8-1863dd40695c" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.543s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.874862] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "4146c49d-fd4b-49f6-8b17-76d06fd143af" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.543s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.991120] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333798, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.240631] env[61906]: DEBUG nova.compute.manager [req-a8a03fb2-63ce-4e81-8723-3163ac9652e8 req-3a26ba16-d281-40c0-bcc2-854d9b905547 service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Received event network-changed-8c2c45ef-6d72-4a38-bce1-9cbc751f3274 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1156.240919] env[61906]: DEBUG nova.compute.manager [req-a8a03fb2-63ce-4e81-8723-3163ac9652e8 req-3a26ba16-d281-40c0-bcc2-854d9b905547 service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Refreshing instance network info cache due to event network-changed-8c2c45ef-6d72-4a38-bce1-9cbc751f3274. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1156.241092] env[61906]: DEBUG oslo_concurrency.lockutils [req-a8a03fb2-63ce-4e81-8723-3163ac9652e8 req-3a26ba16-d281-40c0-bcc2-854d9b905547 service nova] Acquiring lock "refresh_cache-d4a6abdd-75ee-4040-8ca8-1863dd40695c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1156.241257] env[61906]: DEBUG oslo_concurrency.lockutils [req-a8a03fb2-63ce-4e81-8723-3163ac9652e8 req-3a26ba16-d281-40c0-bcc2-854d9b905547 service nova] Acquired lock "refresh_cache-d4a6abdd-75ee-4040-8ca8-1863dd40695c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.241427] env[61906]: DEBUG nova.network.neutron [req-a8a03fb2-63ce-4e81-8723-3163ac9652e8 req-3a26ba16-d281-40c0-bcc2-854d9b905547 service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Refreshing network info cache for port 8c2c45ef-6d72-4a38-bce1-9cbc751f3274 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1156.492587] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333798, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.935908] env[61906]: DEBUG nova.network.neutron [req-a8a03fb2-63ce-4e81-8723-3163ac9652e8 req-3a26ba16-d281-40c0-bcc2-854d9b905547 service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Updated VIF entry in instance network info cache for port 8c2c45ef-6d72-4a38-bce1-9cbc751f3274. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1156.936325] env[61906]: DEBUG nova.network.neutron [req-a8a03fb2-63ce-4e81-8723-3163ac9652e8 req-3a26ba16-d281-40c0-bcc2-854d9b905547 service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Updating instance_info_cache with network_info: [{"id": "8c2c45ef-6d72-4a38-bce1-9cbc751f3274", "address": "fa:16:3e:5c:a3:d4", "network": {"id": "9903d8f6-78ef-4ec6-8386-c031ceb715f6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1583036471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d6fb2749df6498f82aaf7c5529b309f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c2c45ef-6d", "ovs_interfaceid": "8c2c45ef-6d72-4a38-bce1-9cbc751f3274", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1156.993307] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333798, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.188315} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.993561] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f6304fcd-34fc-4c19-a45b-e5229f8ce75a/f6304fcd-34fc-4c19-a45b-e5229f8ce75a.vmdk to [datastore1] 151cc6f4-96c5-4ac3-97d5-73cddfa7b465/151cc6f4-96c5-4ac3-97d5-73cddfa7b465.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1156.994369] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5a6d623-0116-4851-8387-d50dcbad9270 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.015281] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] 151cc6f4-96c5-4ac3-97d5-73cddfa7b465/151cc6f4-96c5-4ac3-97d5-73cddfa7b465.vmdk or device None with type streamOptimized {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1157.015754] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b4bd374-74ff-4cb5-b041-68bcf82f6902 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.034056] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1157.034056] env[61906]: value = "task-1333799" [ 1157.034056] env[61906]: _type = "Task" [ 1157.034056] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.041249] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333799, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.439714] env[61906]: DEBUG oslo_concurrency.lockutils [req-a8a03fb2-63ce-4e81-8723-3163ac9652e8 req-3a26ba16-d281-40c0-bcc2-854d9b905547 service nova] Releasing lock "refresh_cache-d4a6abdd-75ee-4040-8ca8-1863dd40695c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1157.544151] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333799, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.044829] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333799, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.349574] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ebf1ab63-e6c3-4d61-bd4c-dbb10d053055 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "4146c49d-fd4b-49f6-8b17-76d06fd143af" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.350243] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ebf1ab63-e6c3-4d61-bd4c-dbb10d053055 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "4146c49d-fd4b-49f6-8b17-76d06fd143af" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.545389] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333799, 'name': ReconfigVM_Task, 'duration_secs': 1.419815} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.545754] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Reconfigured VM instance instance-00000068 to attach disk [datastore1] 151cc6f4-96c5-4ac3-97d5-73cddfa7b465/151cc6f4-96c5-4ac3-97d5-73cddfa7b465.vmdk or device None with type streamOptimized {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1158.547054] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'device_type': 'disk', 'encrypted': False, 'encryption_secret_uuid': None, 'boot_index': 0, 'disk_bus': None, 'size': 0, 'encryption_options': None, 'guest_format': None, 'encryption_format': None, 'device_name': '/dev/sda', 'image_id': 'a9502563-1f96-46a5-b87e-12c0e9375e6e'}], 'ephemerals': [], 'block_device_mapping': [{'device_type': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284850', 'volume_id': '3e22fc47-57b1-4c7b-b334-e0e05317199f', 'name': 'volume-3e22fc47-57b1-4c7b-b334-e0e05317199f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '151cc6f4-96c5-4ac3-97d5-73cddfa7b465', 'attached_at': '', 'detached_at': '', 'volume_id': '3e22fc47-57b1-4c7b-b334-e0e05317199f', 'serial': '3e22fc47-57b1-4c7b-b334-e0e05317199f'}, 'attachment_id': 'a0519a8c-cc28-4f0f-9cd8-28ee1f3b1a8a', 'delete_on_termination': False, 'boot_index': None, 'disk_bus': None, 'guest_format': None, 'mount_device': '/dev/sdb', 'volume_type': None}], 'swap': None} {{(pid=61906) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1158.547273] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Volume attach. Driver type: vmdk {{(pid=61906) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1158.547485] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284850', 'volume_id': '3e22fc47-57b1-4c7b-b334-e0e05317199f', 'name': 'volume-3e22fc47-57b1-4c7b-b334-e0e05317199f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '151cc6f4-96c5-4ac3-97d5-73cddfa7b465', 'attached_at': '', 'detached_at': '', 'volume_id': '3e22fc47-57b1-4c7b-b334-e0e05317199f', 'serial': '3e22fc47-57b1-4c7b-b334-e0e05317199f'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1158.548282] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ecb5d51-30f8-4eb8-a332-43d0e022f096 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.563438] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80ac97cf-2890-4f83-a1ef-83f22b0b9225 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.586444] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] volume-3e22fc47-57b1-4c7b-b334-e0e05317199f/volume-3e22fc47-57b1-4c7b-b334-e0e05317199f.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1158.586675] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e3bbfd69-3bb6-44c2-a50c-f29d32bff3ad {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.606206] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1158.606206] env[61906]: value = "task-1333800" [ 1158.606206] env[61906]: _type = "Task" [ 1158.606206] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.615180] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333800, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.853434] env[61906]: DEBUG nova.compute.utils [None req-ebf1ab63-e6c3-4d61-bd4c-dbb10d053055 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1159.116149] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333800, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.356624] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ebf1ab63-e6c3-4d61-bd4c-dbb10d053055 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "4146c49d-fd4b-49f6-8b17-76d06fd143af" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1159.616761] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333800, 'name': ReconfigVM_Task, 'duration_secs': 1.005581} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.617093] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Reconfigured VM instance instance-00000068 to attach disk [datastore1] volume-3e22fc47-57b1-4c7b-b334-e0e05317199f/volume-3e22fc47-57b1-4c7b-b334-e0e05317199f.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1159.621886] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-42e7e22b-24a8-48ca-8a7e-d81294213482 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.635920] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1159.635920] env[61906]: value = "task-1333801" [ 1159.635920] env[61906]: _type = "Task" [ 1159.635920] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.643535] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333801, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.146442] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333801, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.414232] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ebf1ab63-e6c3-4d61-bd4c-dbb10d053055 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "4146c49d-fd4b-49f6-8b17-76d06fd143af" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1160.414513] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ebf1ab63-e6c3-4d61-bd4c-dbb10d053055 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "4146c49d-fd4b-49f6-8b17-76d06fd143af" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1160.414761] env[61906]: INFO nova.compute.manager [None req-ebf1ab63-e6c3-4d61-bd4c-dbb10d053055 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Attaching volume 0095d3c9-ba6c-478e-852c-a6f285a79864 to /dev/sdb [ 1160.446273] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37d9a967-4f67-4efb-8a67-cc2160982b0f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.453916] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d1f2aa-c837-4f51-81fb-4fc3d4b66525 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.467781] env[61906]: DEBUG nova.virt.block_device [None req-ebf1ab63-e6c3-4d61-bd4c-dbb10d053055 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Updating existing volume attachment record: aa99e0e3-1ea0-4f59-874e-ee158fff01bf {{(pid=61906) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1160.648212] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333801, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.147880] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333801, 'name': ReconfigVM_Task, 'duration_secs': 1.081788} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.148220] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284850', 'volume_id': '3e22fc47-57b1-4c7b-b334-e0e05317199f', 'name': 'volume-3e22fc47-57b1-4c7b-b334-e0e05317199f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '151cc6f4-96c5-4ac3-97d5-73cddfa7b465', 'attached_at': '', 'detached_at': '', 'volume_id': '3e22fc47-57b1-4c7b-b334-e0e05317199f', 'serial': '3e22fc47-57b1-4c7b-b334-e0e05317199f'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1161.148790] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-313a28b6-ef02-4ab4-971d-4cafa37b6ce4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.154846] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1161.154846] env[61906]: value = "task-1333803" [ 1161.154846] env[61906]: _type = "Task" [ 1161.154846] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.162498] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333803, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.664460] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333803, 'name': Rename_Task, 'duration_secs': 0.137801} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.664798] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1161.664924] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-94e9dda3-5932-489b-b3c3-65f6caa061b0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.672308] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1161.672308] env[61906]: value = "task-1333804" [ 1161.672308] env[61906]: _type = "Task" [ 1161.672308] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.680685] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333804, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.184112] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333804, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.683020] env[61906]: DEBUG oslo_vmware.api [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333804, 'name': PowerOnVM_Task, 'duration_secs': 0.597406} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.683020] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1162.781167] env[61906]: DEBUG nova.compute.manager [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1162.782160] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f001b3e2-30d6-4a93-a1ff-1b53b0a02862 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.300054] env[61906]: DEBUG oslo_concurrency.lockutils [None req-862ca25e-f9d3-473c-bb9f-0fe42ad9c40d tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 29.173s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.300916] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 7.970s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.301273] env[61906]: INFO nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] During sync_power_state the instance has a pending task (spawning). Skip. [ 1163.301497] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.017279] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebf1ab63-e6c3-4d61-bd4c-dbb10d053055 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Volume attach. Driver type: vmdk {{(pid=61906) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1165.017565] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebf1ab63-e6c3-4d61-bd4c-dbb10d053055 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284853', 'volume_id': '0095d3c9-ba6c-478e-852c-a6f285a79864', 'name': 'volume-0095d3c9-ba6c-478e-852c-a6f285a79864', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4146c49d-fd4b-49f6-8b17-76d06fd143af', 'attached_at': '', 'detached_at': '', 'volume_id': '0095d3c9-ba6c-478e-852c-a6f285a79864', 'serial': '0095d3c9-ba6c-478e-852c-a6f285a79864'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1165.018484] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bfe18ae-3c15-439f-af6f-c081a97e2f7b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.034874] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-770f4881-62e5-4c78-b223-e50ba95fffc1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.058497] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebf1ab63-e6c3-4d61-bd4c-dbb10d053055 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] volume-0095d3c9-ba6c-478e-852c-a6f285a79864/volume-0095d3c9-ba6c-478e-852c-a6f285a79864.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1165.058762] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb51d3dd-1d9f-4bcf-b792-988cdd91cc5e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.075633] env[61906]: DEBUG oslo_vmware.api [None req-ebf1ab63-e6c3-4d61-bd4c-dbb10d053055 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1165.075633] env[61906]: value = "task-1333806" [ 1165.075633] env[61906]: _type = "Task" [ 1165.075633] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.083063] env[61906]: DEBUG oslo_vmware.api [None req-ebf1ab63-e6c3-4d61-bd4c-dbb10d053055 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333806, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.586339] env[61906]: DEBUG oslo_vmware.api [None req-ebf1ab63-e6c3-4d61-bd4c-dbb10d053055 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333806, 'name': ReconfigVM_Task, 'duration_secs': 0.3138} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.586639] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebf1ab63-e6c3-4d61-bd4c-dbb10d053055 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Reconfigured VM instance instance-0000006a to attach disk [datastore2] volume-0095d3c9-ba6c-478e-852c-a6f285a79864/volume-0095d3c9-ba6c-478e-852c-a6f285a79864.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1165.591541] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9700029-78e9-4f43-b24e-88028e1c9db3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.606055] env[61906]: DEBUG oslo_vmware.api [None req-ebf1ab63-e6c3-4d61-bd4c-dbb10d053055 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1165.606055] env[61906]: value = "task-1333807" [ 1165.606055] env[61906]: _type = "Task" [ 1165.606055] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.615135] env[61906]: DEBUG oslo_vmware.api [None req-ebf1ab63-e6c3-4d61-bd4c-dbb10d053055 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333807, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.763474] env[61906]: DEBUG nova.compute.manager [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Stashing vm_state: active {{(pid=61906) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1166.116214] env[61906]: DEBUG oslo_vmware.api [None req-ebf1ab63-e6c3-4d61-bd4c-dbb10d053055 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333807, 'name': ReconfigVM_Task, 'duration_secs': 0.124324} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.116665] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebf1ab63-e6c3-4d61-bd4c-dbb10d053055 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284853', 'volume_id': '0095d3c9-ba6c-478e-852c-a6f285a79864', 'name': 'volume-0095d3c9-ba6c-478e-852c-a6f285a79864', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4146c49d-fd4b-49f6-8b17-76d06fd143af', 'attached_at': '', 'detached_at': '', 'volume_id': '0095d3c9-ba6c-478e-852c-a6f285a79864', 'serial': '0095d3c9-ba6c-478e-852c-a6f285a79864'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1166.282675] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1166.282955] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1166.787899] env[61906]: INFO nova.compute.claims [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1167.152044] env[61906]: DEBUG nova.objects.instance [None req-ebf1ab63-e6c3-4d61-bd4c-dbb10d053055 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lazy-loading 'flavor' on Instance uuid 4146c49d-fd4b-49f6-8b17-76d06fd143af {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1167.293964] env[61906]: INFO nova.compute.resource_tracker [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Updating resource usage from migration 59932ae5-ef23-4b83-ad2d-3ce5808f488f [ 1167.387516] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b273e6f1-9b4c-4765-9f98-5343aebb0236 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.395784] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bd7a0a9-feba-4026-b541-6680c8914700 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.426201] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df2b8342-01ba-470d-ba9e-cfa640496317 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.434230] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11803d35-98da-4a35-bd2a-491e31adbabe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.447058] env[61906]: DEBUG nova.compute.provider_tree [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1167.657259] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ebf1ab63-e6c3-4d61-bd4c-dbb10d053055 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "4146c49d-fd4b-49f6-8b17-76d06fd143af" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.243s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.861779] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a08c7ce6-fc14-4510-9253-e05aee1520ae tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "4146c49d-fd4b-49f6-8b17-76d06fd143af" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1167.862054] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a08c7ce6-fc14-4510-9253-e05aee1520ae tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "4146c49d-fd4b-49f6-8b17-76d06fd143af" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1167.907967] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1167.908206] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1167.908304] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Starting heal instance info cache {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1167.949884] env[61906]: DEBUG nova.scheduler.client.report [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1168.364876] env[61906]: INFO nova.compute.manager [None req-a08c7ce6-fc14-4510-9253-e05aee1520ae tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Detaching volume 0095d3c9-ba6c-478e-852c-a6f285a79864 [ 1168.397056] env[61906]: INFO nova.virt.block_device [None req-a08c7ce6-fc14-4510-9253-e05aee1520ae tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Attempting to driver detach volume 0095d3c9-ba6c-478e-852c-a6f285a79864 from mountpoint /dev/sdb [ 1168.397056] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-a08c7ce6-fc14-4510-9253-e05aee1520ae tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Volume detach. Driver type: vmdk {{(pid=61906) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1168.397056] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-a08c7ce6-fc14-4510-9253-e05aee1520ae tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284853', 'volume_id': '0095d3c9-ba6c-478e-852c-a6f285a79864', 'name': 'volume-0095d3c9-ba6c-478e-852c-a6f285a79864', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4146c49d-fd4b-49f6-8b17-76d06fd143af', 'attached_at': '', 'detached_at': '', 'volume_id': '0095d3c9-ba6c-478e-852c-a6f285a79864', 'serial': '0095d3c9-ba6c-478e-852c-a6f285a79864'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1168.397322] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc331912-3bcd-4880-b6f7-80e36002a1fe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.421032] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08438311-439c-487a-8b06-e620abf8598a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.428053] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d66427b3-1151-4080-943e-439ef0ab7f0c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.448427] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "refresh_cache-f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1168.448556] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquired lock "refresh_cache-f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1168.448701] env[61906]: DEBUG nova.network.neutron [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Forcefully refreshing network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1168.450227] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ff27bcd-62b2-45cf-a82c-09135634d7d5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.454581] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.172s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.454772] env[61906]: INFO nova.compute.manager [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Migrating [ 1168.471582] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-a08c7ce6-fc14-4510-9253-e05aee1520ae tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] The volume has not been displaced from its original location: [datastore2] volume-0095d3c9-ba6c-478e-852c-a6f285a79864/volume-0095d3c9-ba6c-478e-852c-a6f285a79864.vmdk. No consolidation needed. {{(pid=61906) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1168.476695] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-a08c7ce6-fc14-4510-9253-e05aee1520ae tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Reconfiguring VM instance instance-0000006a to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1168.479381] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1242cffd-c3fe-4d5f-b23c-217b49c4bdec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.500822] env[61906]: DEBUG oslo_vmware.api [None req-a08c7ce6-fc14-4510-9253-e05aee1520ae tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1168.500822] env[61906]: value = "task-1333808" [ 1168.500822] env[61906]: _type = "Task" [ 1168.500822] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.509631] env[61906]: DEBUG oslo_vmware.api [None req-a08c7ce6-fc14-4510-9253-e05aee1520ae tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333808, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.996661] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "refresh_cache-e75681a9-c315-4f60-9095-fa60e7e39c22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1168.996843] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired lock "refresh_cache-e75681a9-c315-4f60-9095-fa60e7e39c22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1168.997075] env[61906]: DEBUG nova.network.neutron [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1169.011448] env[61906]: DEBUG oslo_vmware.api [None req-a08c7ce6-fc14-4510-9253-e05aee1520ae tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333808, 'name': ReconfigVM_Task, 'duration_secs': 0.213434} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.011702] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-a08c7ce6-fc14-4510-9253-e05aee1520ae tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Reconfigured VM instance instance-0000006a to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1169.016756] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-149c377a-68e6-482e-b7dc-2ea6f633b747 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.030728] env[61906]: DEBUG oslo_vmware.api [None req-a08c7ce6-fc14-4510-9253-e05aee1520ae tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1169.030728] env[61906]: value = "task-1333809" [ 1169.030728] env[61906]: _type = "Task" [ 1169.030728] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.039193] env[61906]: DEBUG oslo_vmware.api [None req-a08c7ce6-fc14-4510-9253-e05aee1520ae tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333809, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.541299] env[61906]: DEBUG oslo_vmware.api [None req-a08c7ce6-fc14-4510-9253-e05aee1520ae tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333809, 'name': ReconfigVM_Task, 'duration_secs': 0.124426} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.541589] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-a08c7ce6-fc14-4510-9253-e05aee1520ae tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284853', 'volume_id': '0095d3c9-ba6c-478e-852c-a6f285a79864', 'name': 'volume-0095d3c9-ba6c-478e-852c-a6f285a79864', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4146c49d-fd4b-49f6-8b17-76d06fd143af', 'attached_at': '', 'detached_at': '', 'volume_id': '0095d3c9-ba6c-478e-852c-a6f285a79864', 'serial': '0095d3c9-ba6c-478e-852c-a6f285a79864'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1169.658400] env[61906]: DEBUG nova.network.neutron [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Updating instance_info_cache with network_info: [{"id": "697dbd16-ab06-44ba-ab39-8da5e1ce0c78", "address": "fa:16:3e:80:02:ba", "network": {"id": "9903d8f6-78ef-4ec6-8386-c031ceb715f6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1583036471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d6fb2749df6498f82aaf7c5529b309f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap697dbd16-ab", "ovs_interfaceid": "697dbd16-ab06-44ba-ab39-8da5e1ce0c78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1169.720481] env[61906]: DEBUG nova.network.neutron [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Updating instance_info_cache with network_info: [{"id": "1b29cb2d-b70e-4022-b7c7-fd014b216c1c", "address": "fa:16:3e:7f:a9:4c", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b29cb2d-b7", "ovs_interfaceid": "1b29cb2d-b70e-4022-b7c7-fd014b216c1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1170.081459] env[61906]: DEBUG nova.objects.instance [None req-a08c7ce6-fc14-4510-9253-e05aee1520ae tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lazy-loading 'flavor' on Instance uuid 4146c49d-fd4b-49f6-8b17-76d06fd143af {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1170.160828] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Releasing lock "refresh_cache-f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1170.161053] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Updated the network info_cache for instance {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1170.161256] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1170.161413] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1170.161564] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1170.161707] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1170.161833] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61906) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1170.161975] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1170.223570] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Releasing lock "refresh_cache-e75681a9-c315-4f60-9095-fa60e7e39c22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1170.664363] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.664705] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.664705] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.664871] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61906) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1170.665767] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5844a914-7ade-4e91-bd82-5256351ed53f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.673557] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edadc32a-f301-459e-9024-ca13420b0657 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.686899] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49bcde4c-48f1-4358-b9ed-1120ef26551b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.693013] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0b13d98-33bc-4c63-8e1a-60fda3e4b51a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.720972] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180106MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=61906) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1170.721132] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.721323] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1171.088276] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a08c7ce6-fc14-4510-9253-e05aee1520ae tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "4146c49d-fd4b-49f6-8b17-76d06fd143af" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.226s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.729163] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Applying migration context for instance e75681a9-c315-4f60-9095-fa60e7e39c22 as it has an incoming, in-progress migration 59932ae5-ef23-4b83-ad2d-3ce5808f488f. Migration status is migrating {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1171.730118] env[61906]: INFO nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Updating resource usage from migration 59932ae5-ef23-4b83-ad2d-3ce5808f488f [ 1171.735563] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9390223-b3a9-4fc9-a380-3b4273075bd4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.753516] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Updating instance 'e75681a9-c315-4f60-9095-fa60e7e39c22' progress to 0 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1171.757281] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 86be7197-c0d1-49c5-aa30-cf908a506031 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1171.757430] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1171.757567] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance d4a6abdd-75ee-4040-8ca8-1863dd40695c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1171.757686] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 4146c49d-fd4b-49f6-8b17-76d06fd143af actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1171.757803] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 151cc6f4-96c5-4ac3-97d5-73cddfa7b465 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1171.757916] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Migration 59932ae5-ef23-4b83-ad2d-3ce5808f488f is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1171.758039] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance e75681a9-c315-4f60-9095-fa60e7e39c22 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1171.758239] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1171.758387] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1920MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1171.838204] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6f94b52-ee77-454e-af56-e1572c26bf1f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.845863] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d055f6-803c-4391-a6fd-261fd7b31947 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.874375] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7404e29c-7cfc-4ebb-9267-6042fe6966ba {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.880987] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5fd9084-7fdd-4bc2-aa02-94ed5a42b754 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.893274] env[61906]: DEBUG nova.compute.provider_tree [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1172.092355] env[61906]: DEBUG oslo_concurrency.lockutils [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "4146c49d-fd4b-49f6-8b17-76d06fd143af" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1172.092637] env[61906]: DEBUG oslo_concurrency.lockutils [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "4146c49d-fd4b-49f6-8b17-76d06fd143af" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1172.092868] env[61906]: DEBUG oslo_concurrency.lockutils [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "4146c49d-fd4b-49f6-8b17-76d06fd143af-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1172.093061] env[61906]: DEBUG oslo_concurrency.lockutils [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "4146c49d-fd4b-49f6-8b17-76d06fd143af-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1172.093239] env[61906]: DEBUG oslo_concurrency.lockutils [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "4146c49d-fd4b-49f6-8b17-76d06fd143af-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1172.095761] env[61906]: INFO nova.compute.manager [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Terminating instance [ 1172.098183] env[61906]: DEBUG nova.compute.manager [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1172.098362] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1172.099202] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a864e5dd-47db-4326-aaeb-8cc5d05eca5c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.108419] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1172.108648] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ab27dd63-4cec-40a5-8b5c-fa88e1d1d8b4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.115558] env[61906]: DEBUG oslo_vmware.api [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1172.115558] env[61906]: value = "task-1333810" [ 1172.115558] env[61906]: _type = "Task" [ 1172.115558] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.123308] env[61906]: DEBUG oslo_vmware.api [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333810, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.263087] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1172.263500] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7571a8b1-67ad-474e-905d-6322f59a672a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.270663] env[61906]: DEBUG oslo_vmware.api [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1172.270663] env[61906]: value = "task-1333811" [ 1172.270663] env[61906]: _type = "Task" [ 1172.270663] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.279108] env[61906]: DEBUG oslo_vmware.api [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333811, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.396426] env[61906]: DEBUG nova.scheduler.client.report [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1172.625302] env[61906]: DEBUG oslo_vmware.api [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333810, 'name': PowerOffVM_Task, 'duration_secs': 0.179496} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.625552] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1172.625724] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1172.625974] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-af2f7ae2-2cf2-4226-b61b-ea07010f8ad9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.683459] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1172.683697] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1172.683858] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Deleting the datastore file [datastore1] 4146c49d-fd4b-49f6-8b17-76d06fd143af {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1172.684130] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cdbfb278-dea5-4bb4-873c-eb8b54acd48d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.690144] env[61906]: DEBUG oslo_vmware.api [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for the task: (returnval){ [ 1172.690144] env[61906]: value = "task-1333813" [ 1172.690144] env[61906]: _type = "Task" [ 1172.690144] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.697251] env[61906]: DEBUG oslo_vmware.api [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333813, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.780658] env[61906]: DEBUG oslo_vmware.api [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333811, 'name': PowerOffVM_Task, 'duration_secs': 0.169172} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.781014] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1172.781256] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Updating instance 'e75681a9-c315-4f60-9095-fa60e7e39c22' progress to 17 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1172.901075] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1172.901280] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.180s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1173.125306] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1173.125532] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1173.200253] env[61906]: DEBUG oslo_vmware.api [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Task: {'id': task-1333813, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130149} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.200633] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1173.200878] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1173.201103] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1173.201301] env[61906]: INFO nova.compute.manager [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1173.201553] env[61906]: DEBUG oslo.service.loopingcall [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1173.201752] env[61906]: DEBUG nova.compute.manager [-] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1173.201849] env[61906]: DEBUG nova.network.neutron [-] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1173.288358] env[61906]: DEBUG nova.virt.hardware [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:23Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1173.288622] env[61906]: DEBUG nova.virt.hardware [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1173.288766] env[61906]: DEBUG nova.virt.hardware [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1173.288954] env[61906]: DEBUG nova.virt.hardware [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1173.289122] env[61906]: DEBUG nova.virt.hardware [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1173.289284] env[61906]: DEBUG nova.virt.hardware [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1173.289491] env[61906]: DEBUG nova.virt.hardware [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1173.289708] env[61906]: DEBUG nova.virt.hardware [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1173.289822] env[61906]: DEBUG nova.virt.hardware [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1173.289986] env[61906]: DEBUG nova.virt.hardware [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1173.290200] env[61906]: DEBUG nova.virt.hardware [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1173.295291] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8dcb24f0-7be2-47f7-a324-e7c109821509 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.311830] env[61906]: DEBUG oslo_vmware.api [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1173.311830] env[61906]: value = "task-1333814" [ 1173.311830] env[61906]: _type = "Task" [ 1173.311830] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.321922] env[61906]: DEBUG oslo_vmware.api [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333814, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.643371] env[61906]: DEBUG nova.compute.manager [req-9c6c5bee-d644-49e3-bb86-ae69ac2f7c19 req-eac3e30b-6a80-4f71-bc37-f61a03f35925 service nova] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Received event network-vif-deleted-eaae61e5-a80d-4b75-aaf2-149cb19e3409 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1173.644260] env[61906]: INFO nova.compute.manager [req-9c6c5bee-d644-49e3-bb86-ae69ac2f7c19 req-eac3e30b-6a80-4f71-bc37-f61a03f35925 service nova] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Neutron deleted interface eaae61e5-a80d-4b75-aaf2-149cb19e3409; detaching it from the instance and deleting it from the info cache [ 1173.644260] env[61906]: DEBUG nova.network.neutron [req-9c6c5bee-d644-49e3-bb86-ae69ac2f7c19 req-eac3e30b-6a80-4f71-bc37-f61a03f35925 service nova] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1173.821640] env[61906]: DEBUG oslo_vmware.api [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333814, 'name': ReconfigVM_Task, 'duration_secs': 0.156873} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.822009] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Updating instance 'e75681a9-c315-4f60-9095-fa60e7e39c22' progress to 33 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1174.121780] env[61906]: DEBUG nova.network.neutron [-] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1174.146348] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-42e8aa9b-5b3d-4880-a06d-07b59a23448d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.158217] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f3c7c4d-4cad-4d0b-8909-ad01ee7bd327 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.184360] env[61906]: DEBUG nova.compute.manager [req-9c6c5bee-d644-49e3-bb86-ae69ac2f7c19 req-eac3e30b-6a80-4f71-bc37-f61a03f35925 service nova] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Detach interface failed, port_id=eaae61e5-a80d-4b75-aaf2-149cb19e3409, reason: Instance 4146c49d-fd4b-49f6-8b17-76d06fd143af could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1174.328793] env[61906]: DEBUG nova.virt.hardware [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1174.329072] env[61906]: DEBUG nova.virt.hardware [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1174.329246] env[61906]: DEBUG nova.virt.hardware [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1174.329439] env[61906]: DEBUG nova.virt.hardware [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1174.329594] env[61906]: DEBUG nova.virt.hardware [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1174.329746] env[61906]: DEBUG nova.virt.hardware [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1174.329949] env[61906]: DEBUG nova.virt.hardware [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1174.330137] env[61906]: DEBUG nova.virt.hardware [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1174.330314] env[61906]: DEBUG nova.virt.hardware [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1174.330481] env[61906]: DEBUG nova.virt.hardware [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1174.330657] env[61906]: DEBUG nova.virt.hardware [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1174.335945] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Reconfiguring VM instance instance-0000006b to detach disk 2000 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1174.336405] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8a577118-a481-4ff7-93f3-1f9437df394f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.354611] env[61906]: DEBUG oslo_vmware.api [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1174.354611] env[61906]: value = "task-1333815" [ 1174.354611] env[61906]: _type = "Task" [ 1174.354611] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.362389] env[61906]: DEBUG oslo_vmware.api [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333815, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.625352] env[61906]: INFO nova.compute.manager [-] [instance: 4146c49d-fd4b-49f6-8b17-76d06fd143af] Took 1.42 seconds to deallocate network for instance. [ 1174.863822] env[61906]: DEBUG oslo_vmware.api [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333815, 'name': ReconfigVM_Task, 'duration_secs': 0.148852} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.864186] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Reconfigured VM instance instance-0000006b to detach disk 2000 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1174.864870] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d88bb656-a1cd-4949-bcda-268f6b82de51 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.885981] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] e75681a9-c315-4f60-9095-fa60e7e39c22/e75681a9-c315-4f60-9095-fa60e7e39c22.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1174.886246] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dc6b7fb1-e4d4-4b4a-b580-2bd620940a43 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.904063] env[61906]: DEBUG oslo_vmware.api [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1174.904063] env[61906]: value = "task-1333816" [ 1174.904063] env[61906]: _type = "Task" [ 1174.904063] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.911258] env[61906]: DEBUG oslo_vmware.api [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333816, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.131365] env[61906]: DEBUG oslo_concurrency.lockutils [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1175.131601] env[61906]: DEBUG oslo_concurrency.lockutils [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1175.131826] env[61906]: DEBUG nova.objects.instance [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lazy-loading 'resources' on Instance uuid 4146c49d-fd4b-49f6-8b17-76d06fd143af {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1175.413777] env[61906]: DEBUG oslo_vmware.api [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333816, 'name': ReconfigVM_Task, 'duration_secs': 0.250692} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.414016] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Reconfigured VM instance instance-0000006b to attach disk [datastore2] e75681a9-c315-4f60-9095-fa60e7e39c22/e75681a9-c315-4f60-9095-fa60e7e39c22.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1175.414303] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Updating instance 'e75681a9-c315-4f60-9095-fa60e7e39c22' progress to 50 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1175.731286] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0231f8bf-f86c-4d16-9a7f-a0d2ed70c861 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.739089] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5fae056-0cb6-4133-8f03-4c0b72e1a52e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.770462] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbf992da-b602-470a-bde2-494a9d1200c6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.777391] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-551bbcb4-207e-43a1-ab4d-4f0aaa38dfbc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.790516] env[61906]: DEBUG nova.compute.provider_tree [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1175.920373] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96c39046-972a-42f7-92e6-3d66d1f1e894 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.942096] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fc1cef2-f456-453b-bc27-6995b1a9d2b8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.967727] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Updating instance 'e75681a9-c315-4f60-9095-fa60e7e39c22' progress to 67 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1176.293563] env[61906]: DEBUG nova.scheduler.client.report [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1176.511461] env[61906]: DEBUG nova.network.neutron [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Port 1b29cb2d-b70e-4022-b7c7-fd014b216c1c binding to destination host cpu-1 is already ACTIVE {{(pid=61906) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1176.798702] env[61906]: DEBUG oslo_concurrency.lockutils [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.667s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1176.819161] env[61906]: INFO nova.scheduler.client.report [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Deleted allocations for instance 4146c49d-fd4b-49f6-8b17-76d06fd143af [ 1177.326473] env[61906]: DEBUG oslo_concurrency.lockutils [None req-18b7873d-2384-433f-91c2-b5dbd63b4137 tempest-AttachVolumeNegativeTest-30355476 tempest-AttachVolumeNegativeTest-30355476-project-member] Lock "4146c49d-fd4b-49f6-8b17-76d06fd143af" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.234s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1177.536053] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "e75681a9-c315-4f60-9095-fa60e7e39c22-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1177.536053] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "e75681a9-c315-4f60-9095-fa60e7e39c22-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1177.536053] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "e75681a9-c315-4f60-9095-fa60e7e39c22-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1178.568670] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "refresh_cache-e75681a9-c315-4f60-9095-fa60e7e39c22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1178.568891] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired lock "refresh_cache-e75681a9-c315-4f60-9095-fa60e7e39c22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.569064] env[61906]: DEBUG nova.network.neutron [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1179.282136] env[61906]: DEBUG nova.network.neutron [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Updating instance_info_cache with network_info: [{"id": "1b29cb2d-b70e-4022-b7c7-fd014b216c1c", "address": "fa:16:3e:7f:a9:4c", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b29cb2d-b7", "ovs_interfaceid": "1b29cb2d-b70e-4022-b7c7-fd014b216c1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1179.784724] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Releasing lock "refresh_cache-e75681a9-c315-4f60-9095-fa60e7e39c22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1180.309707] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ccee219-7d3e-4854-ba40-4e6885c38b71 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.328237] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19869051-5121-4d16-b663-08c7a510ea48 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.334474] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Updating instance 'e75681a9-c315-4f60-9095-fa60e7e39c22' progress to 83 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1180.841062] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1180.841416] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b2ded88c-a6b1-43b9-ab82-a9f14f983299 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.848481] env[61906]: DEBUG oslo_vmware.api [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1180.848481] env[61906]: value = "task-1333819" [ 1180.848481] env[61906]: _type = "Task" [ 1180.848481] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.857664] env[61906]: DEBUG oslo_vmware.api [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333819, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.358372] env[61906]: DEBUG oslo_vmware.api [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333819, 'name': PowerOnVM_Task, 'duration_secs': 0.371678} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.358645] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1181.358837] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fb174fee-ccdb-4fe4-92a7-3fe5fc40ebe6 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Updating instance 'e75681a9-c315-4f60-9095-fa60e7e39c22' progress to 100 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1183.177817] env[61906]: DEBUG nova.network.neutron [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Port 1b29cb2d-b70e-4022-b7c7-fd014b216c1c binding to destination host cpu-1 is already ACTIVE {{(pid=61906) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1183.178124] env[61906]: DEBUG oslo_concurrency.lockutils [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "refresh_cache-e75681a9-c315-4f60-9095-fa60e7e39c22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1183.178264] env[61906]: DEBUG oslo_concurrency.lockutils [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired lock "refresh_cache-e75681a9-c315-4f60-9095-fa60e7e39c22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.178431] env[61906]: DEBUG nova.network.neutron [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1183.933065] env[61906]: DEBUG nova.network.neutron [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Updating instance_info_cache with network_info: [{"id": "1b29cb2d-b70e-4022-b7c7-fd014b216c1c", "address": "fa:16:3e:7f:a9:4c", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b29cb2d-b7", "ovs_interfaceid": "1b29cb2d-b70e-4022-b7c7-fd014b216c1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1184.436341] env[61906]: DEBUG oslo_concurrency.lockutils [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Releasing lock "refresh_cache-e75681a9-c315-4f60-9095-fa60e7e39c22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1184.940863] env[61906]: DEBUG nova.compute.manager [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61906) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1184.941175] env[61906]: DEBUG oslo_concurrency.lockutils [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.941444] env[61906]: DEBUG oslo_concurrency.lockutils [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.444825] env[61906]: DEBUG nova.objects.instance [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lazy-loading 'migration_context' on Instance uuid e75681a9-c315-4f60-9095-fa60e7e39c22 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1186.032820] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64645d78-dcc5-41d5-b692-2da791f98d6f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.040144] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b73d3b9c-26e8-4d28-b4b3-991197a29850 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.069087] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c6207dc-9422-45b3-809c-1c7ae367835a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.075787] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49dcc9df-f916-4d47-9210-33a9e89e0fc2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.088497] env[61906]: DEBUG nova.compute.provider_tree [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1186.592359] env[61906]: DEBUG nova.scheduler.client.report [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1187.607984] env[61906]: DEBUG oslo_concurrency.lockutils [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.666s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1189.150497] env[61906]: INFO nova.compute.manager [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Swapping old allocation on dict_keys(['a5b57df7-a16e-44f9-9b2d-23c518860263']) held by migration 59932ae5-ef23-4b83-ad2d-3ce5808f488f for instance [ 1189.173757] env[61906]: DEBUG nova.scheduler.client.report [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Overwriting current allocation {'allocations': {'a5b57df7-a16e-44f9-9b2d-23c518860263': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 137}}, 'project_id': 'b7d3aa88849f4492aa16ff337b8cfdda', 'user_id': '27ba0d3b00ce4513a03259970a600190', 'consumer_generation': 1} on consumer e75681a9-c315-4f60-9095-fa60e7e39c22 {{(pid=61906) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1189.246216] env[61906]: DEBUG oslo_concurrency.lockutils [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "refresh_cache-e75681a9-c315-4f60-9095-fa60e7e39c22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1189.246438] env[61906]: DEBUG oslo_concurrency.lockutils [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired lock "refresh_cache-e75681a9-c315-4f60-9095-fa60e7e39c22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.246733] env[61906]: DEBUG nova.network.neutron [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1189.974011] env[61906]: DEBUG nova.network.neutron [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Updating instance_info_cache with network_info: [{"id": "1b29cb2d-b70e-4022-b7c7-fd014b216c1c", "address": "fa:16:3e:7f:a9:4c", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b29cb2d-b7", "ovs_interfaceid": "1b29cb2d-b70e-4022-b7c7-fd014b216c1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1190.055672] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4902507d-3825-4272-ba39-f9ab5538a260 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "d4a6abdd-75ee-4040-8ca8-1863dd40695c" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1190.055952] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4902507d-3825-4272-ba39-f9ab5538a260 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "d4a6abdd-75ee-4040-8ca8-1863dd40695c" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1190.477384] env[61906]: DEBUG oslo_concurrency.lockutils [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Releasing lock "refresh_cache-e75681a9-c315-4f60-9095-fa60e7e39c22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1190.477806] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1190.478116] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c163b75f-bb9f-4492-9189-5141ff9f6012 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.486021] env[61906]: DEBUG oslo_vmware.api [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1190.486021] env[61906]: value = "task-1333821" [ 1190.486021] env[61906]: _type = "Task" [ 1190.486021] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.494737] env[61906]: DEBUG oslo_vmware.api [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333821, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.558827] env[61906]: INFO nova.compute.manager [None req-4902507d-3825-4272-ba39-f9ab5538a260 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Detaching volume cd58e0dc-6ca1-45b1-94fd-4b87651291e9 [ 1190.594256] env[61906]: INFO nova.virt.block_device [None req-4902507d-3825-4272-ba39-f9ab5538a260 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Attempting to driver detach volume cd58e0dc-6ca1-45b1-94fd-4b87651291e9 from mountpoint /dev/sdb [ 1190.594256] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-4902507d-3825-4272-ba39-f9ab5538a260 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Volume detach. Driver type: vmdk {{(pid=61906) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1190.594256] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-4902507d-3825-4272-ba39-f9ab5538a260 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284848', 'volume_id': 'cd58e0dc-6ca1-45b1-94fd-4b87651291e9', 'name': 'volume-cd58e0dc-6ca1-45b1-94fd-4b87651291e9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd4a6abdd-75ee-4040-8ca8-1863dd40695c', 'attached_at': '', 'detached_at': '', 'volume_id': 'cd58e0dc-6ca1-45b1-94fd-4b87651291e9', 'serial': 'cd58e0dc-6ca1-45b1-94fd-4b87651291e9'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1190.595785] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efc56cb5-36fa-4e08-baf0-873136225b57 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.620760] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c933738f-1214-47a4-9b11-dff0e4ba0c4f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.629798] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4097cef-c7fe-432b-88fa-ae5d6b69711f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.652355] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4624a340-ad00-4719-8ac3-b43fcaf6a28a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.667839] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-4902507d-3825-4272-ba39-f9ab5538a260 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] The volume has not been displaced from its original location: [datastore1] volume-cd58e0dc-6ca1-45b1-94fd-4b87651291e9/volume-cd58e0dc-6ca1-45b1-94fd-4b87651291e9.vmdk. No consolidation needed. {{(pid=61906) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1190.673409] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-4902507d-3825-4272-ba39-f9ab5538a260 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Reconfiguring VM instance instance-00000069 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1190.673821] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6614bbb9-4e6e-460b-9781-cf4478834d5f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.692793] env[61906]: DEBUG oslo_vmware.api [None req-4902507d-3825-4272-ba39-f9ab5538a260 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1190.692793] env[61906]: value = "task-1333822" [ 1190.692793] env[61906]: _type = "Task" [ 1190.692793] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.703825] env[61906]: DEBUG oslo_vmware.api [None req-4902507d-3825-4272-ba39-f9ab5538a260 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333822, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.995039] env[61906]: DEBUG oslo_vmware.api [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333821, 'name': PowerOffVM_Task, 'duration_secs': 0.217405} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.995396] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1190.996024] env[61906]: DEBUG nova.virt.hardware [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1190.996243] env[61906]: DEBUG nova.virt.hardware [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1190.996406] env[61906]: DEBUG nova.virt.hardware [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1190.996596] env[61906]: DEBUG nova.virt.hardware [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1190.996747] env[61906]: DEBUG nova.virt.hardware [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1190.996903] env[61906]: DEBUG nova.virt.hardware [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1190.997123] env[61906]: DEBUG nova.virt.hardware [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1190.997294] env[61906]: DEBUG nova.virt.hardware [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1190.997469] env[61906]: DEBUG nova.virt.hardware [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1190.997642] env[61906]: DEBUG nova.virt.hardware [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1190.997854] env[61906]: DEBUG nova.virt.hardware [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1191.002804] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-042cc09d-f163-4f13-a25e-20d3380582b9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.017693] env[61906]: DEBUG oslo_vmware.api [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1191.017693] env[61906]: value = "task-1333823" [ 1191.017693] env[61906]: _type = "Task" [ 1191.017693] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.025548] env[61906]: DEBUG oslo_vmware.api [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333823, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.202679] env[61906]: DEBUG oslo_vmware.api [None req-4902507d-3825-4272-ba39-f9ab5538a260 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333822, 'name': ReconfigVM_Task, 'duration_secs': 0.252437} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.202967] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-4902507d-3825-4272-ba39-f9ab5538a260 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Reconfigured VM instance instance-00000069 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1191.207603] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9fc2e7f8-aaa5-4bca-99e5-b3d725b1013d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.222047] env[61906]: DEBUG oslo_vmware.api [None req-4902507d-3825-4272-ba39-f9ab5538a260 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1191.222047] env[61906]: value = "task-1333824" [ 1191.222047] env[61906]: _type = "Task" [ 1191.222047] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.229603] env[61906]: DEBUG oslo_vmware.api [None req-4902507d-3825-4272-ba39-f9ab5538a260 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333824, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.527363] env[61906]: DEBUG oslo_vmware.api [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333823, 'name': ReconfigVM_Task, 'duration_secs': 0.124348} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.528200] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6c3086f-5485-427a-ac75-e01399fa81ad {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.545518] env[61906]: DEBUG nova.virt.hardware [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1191.545759] env[61906]: DEBUG nova.virt.hardware [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1191.545923] env[61906]: DEBUG nova.virt.hardware [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1191.546126] env[61906]: DEBUG nova.virt.hardware [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1191.546282] env[61906]: DEBUG nova.virt.hardware [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1191.546473] env[61906]: DEBUG nova.virt.hardware [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1191.546703] env[61906]: DEBUG nova.virt.hardware [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1191.546869] env[61906]: DEBUG nova.virt.hardware [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1191.547052] env[61906]: DEBUG nova.virt.hardware [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1191.547224] env[61906]: DEBUG nova.virt.hardware [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1191.547400] env[61906]: DEBUG nova.virt.hardware [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1191.548268] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9f51d69-0e85-4b79-837c-a2da06b185bc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.553230] env[61906]: DEBUG oslo_vmware.api [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1191.553230] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52b09d2f-e012-43c2-95d4-f80c59557316" [ 1191.553230] env[61906]: _type = "Task" [ 1191.553230] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.560559] env[61906]: DEBUG oslo_vmware.api [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52b09d2f-e012-43c2-95d4-f80c59557316, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.731088] env[61906]: DEBUG oslo_vmware.api [None req-4902507d-3825-4272-ba39-f9ab5538a260 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333824, 'name': ReconfigVM_Task, 'duration_secs': 0.123311} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.731435] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-4902507d-3825-4272-ba39-f9ab5538a260 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284848', 'volume_id': 'cd58e0dc-6ca1-45b1-94fd-4b87651291e9', 'name': 'volume-cd58e0dc-6ca1-45b1-94fd-4b87651291e9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd4a6abdd-75ee-4040-8ca8-1863dd40695c', 'attached_at': '', 'detached_at': '', 'volume_id': 'cd58e0dc-6ca1-45b1-94fd-4b87651291e9', 'serial': 'cd58e0dc-6ca1-45b1-94fd-4b87651291e9'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1192.064106] env[61906]: DEBUG oslo_vmware.api [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52b09d2f-e012-43c2-95d4-f80c59557316, 'name': SearchDatastore_Task, 'duration_secs': 0.00808} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.069409] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Reconfiguring VM instance instance-0000006b to detach disk 2000 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1192.069682] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d17fa25-a2c7-41c1-8d71-7af5649f9f15 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.086620] env[61906]: DEBUG oslo_vmware.api [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1192.086620] env[61906]: value = "task-1333825" [ 1192.086620] env[61906]: _type = "Task" [ 1192.086620] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.094712] env[61906]: DEBUG oslo_vmware.api [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333825, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.270544] env[61906]: DEBUG nova.objects.instance [None req-4902507d-3825-4272-ba39-f9ab5538a260 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lazy-loading 'flavor' on Instance uuid d4a6abdd-75ee-4040-8ca8-1863dd40695c {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1192.596675] env[61906]: DEBUG oslo_vmware.api [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333825, 'name': ReconfigVM_Task, 'duration_secs': 0.168608} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.597077] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Reconfigured VM instance instance-0000006b to detach disk 2000 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1192.597725] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-795aec8d-d263-4651-8fc4-1c4bd40d67ec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.618798] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] e75681a9-c315-4f60-9095-fa60e7e39c22/e75681a9-c315-4f60-9095-fa60e7e39c22.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1192.619072] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a44ca413-2254-413b-9cc1-7fa70980bbe4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.635642] env[61906]: DEBUG oslo_vmware.api [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1192.635642] env[61906]: value = "task-1333826" [ 1192.635642] env[61906]: _type = "Task" [ 1192.635642] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.643068] env[61906]: DEBUG oslo_vmware.api [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333826, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.145559] env[61906]: DEBUG oslo_vmware.api [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333826, 'name': ReconfigVM_Task, 'duration_secs': 0.240374} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.145854] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Reconfigured VM instance instance-0000006b to attach disk [datastore2] e75681a9-c315-4f60-9095-fa60e7e39c22/e75681a9-c315-4f60-9095-fa60e7e39c22.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1193.146675] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3a847e2-1591-4c38-a520-66f7002b38e7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.163842] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-854a0fa6-936d-4b81-870e-913bf418b4e6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.182870] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-461abfb4-7436-40ff-8ecd-959631f03a11 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.199752] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-196f2911-08c2-496f-8baf-1e4c73825166 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.206064] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1193.206308] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f0f52113-e481-4714-b318-73b14b45cd40 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.211833] env[61906]: DEBUG oslo_vmware.api [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1193.211833] env[61906]: value = "task-1333827" [ 1193.211833] env[61906]: _type = "Task" [ 1193.211833] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.220195] env[61906]: DEBUG oslo_vmware.api [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333827, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.277967] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4902507d-3825-4272-ba39-f9ab5538a260 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "d4a6abdd-75ee-4040-8ca8-1863dd40695c" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.222s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1193.722432] env[61906]: DEBUG oslo_vmware.api [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333827, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.225036] env[61906]: DEBUG oslo_vmware.api [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333827, 'name': PowerOnVM_Task, 'duration_secs': 0.532221} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.225036] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1194.356609] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "d4a6abdd-75ee-4040-8ca8-1863dd40695c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1194.356898] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "d4a6abdd-75ee-4040-8ca8-1863dd40695c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1194.357135] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "d4a6abdd-75ee-4040-8ca8-1863dd40695c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1194.357333] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "d4a6abdd-75ee-4040-8ca8-1863dd40695c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1194.357547] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "d4a6abdd-75ee-4040-8ca8-1863dd40695c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1194.359770] env[61906]: INFO nova.compute.manager [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Terminating instance [ 1194.361522] env[61906]: DEBUG nova.compute.manager [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1194.361720] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1194.362561] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e6db92-9c30-4480-9a5b-7f4db5435f19 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.370602] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1194.371089] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-16955488-31a2-48ad-8048-de1f27111456 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.377120] env[61906]: DEBUG oslo_vmware.api [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1194.377120] env[61906]: value = "task-1333828" [ 1194.377120] env[61906]: _type = "Task" [ 1194.377120] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.384652] env[61906]: DEBUG oslo_vmware.api [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333828, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.887267] env[61906]: DEBUG oslo_vmware.api [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333828, 'name': PowerOffVM_Task, 'duration_secs': 0.166621} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.887711] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1194.887919] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1194.888273] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1b6b831b-3f02-46fd-adfe-16f696c362fe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.943748] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1194.943931] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1194.944167] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Deleting the datastore file [datastore1] d4a6abdd-75ee-4040-8ca8-1863dd40695c {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1194.944448] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5e641ac3-0a0c-4336-97cc-3ceafe03d524 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.950627] env[61906]: DEBUG oslo_vmware.api [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1194.950627] env[61906]: value = "task-1333830" [ 1194.950627] env[61906]: _type = "Task" [ 1194.950627] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.959605] env[61906]: DEBUG oslo_vmware.api [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333830, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.233526] env[61906]: INFO nova.compute.manager [None req-71bf6a7a-f554-4a72-830d-729be0d0f17e tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Updating instance to original state: 'active' [ 1195.461084] env[61906]: DEBUG oslo_vmware.api [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333830, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141648} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.461379] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1195.461570] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1195.461750] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1195.461927] env[61906]: INFO nova.compute.manager [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1195.462212] env[61906]: DEBUG oslo.service.loopingcall [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1195.462414] env[61906]: DEBUG nova.compute.manager [-] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1195.462536] env[61906]: DEBUG nova.network.neutron [-] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1195.965352] env[61906]: DEBUG nova.compute.manager [req-4d294018-dea8-45af-b97a-7d43321d022c req-731b3034-49ad-48b7-a23f-e26ebb0e11dd service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Received event network-vif-deleted-8c2c45ef-6d72-4a38-bce1-9cbc751f3274 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1195.965610] env[61906]: INFO nova.compute.manager [req-4d294018-dea8-45af-b97a-7d43321d022c req-731b3034-49ad-48b7-a23f-e26ebb0e11dd service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Neutron deleted interface 8c2c45ef-6d72-4a38-bce1-9cbc751f3274; detaching it from the instance and deleting it from the info cache [ 1195.965796] env[61906]: DEBUG nova.network.neutron [req-4d294018-dea8-45af-b97a-7d43321d022c req-731b3034-49ad-48b7-a23f-e26ebb0e11dd service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1196.412861] env[61906]: DEBUG nova.network.neutron [-] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1196.468913] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9df3caab-9de6-4746-80c5-fc9a337c199d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.477936] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a5d04a2-8f88-441f-8ce2-81c502f9021b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.502455] env[61906]: DEBUG nova.compute.manager [req-4d294018-dea8-45af-b97a-7d43321d022c req-731b3034-49ad-48b7-a23f-e26ebb0e11dd service nova] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Detach interface failed, port_id=8c2c45ef-6d72-4a38-bce1-9cbc751f3274, reason: Instance d4a6abdd-75ee-4040-8ca8-1863dd40695c could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1196.908044] env[61906]: DEBUG oslo_concurrency.lockutils [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "e75681a9-c315-4f60-9095-fa60e7e39c22" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1196.908270] env[61906]: DEBUG oslo_concurrency.lockutils [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "e75681a9-c315-4f60-9095-fa60e7e39c22" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.908492] env[61906]: DEBUG oslo_concurrency.lockutils [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "e75681a9-c315-4f60-9095-fa60e7e39c22-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1196.908691] env[61906]: DEBUG oslo_concurrency.lockutils [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "e75681a9-c315-4f60-9095-fa60e7e39c22-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.908869] env[61906]: DEBUG oslo_concurrency.lockutils [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "e75681a9-c315-4f60-9095-fa60e7e39c22-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1196.911079] env[61906]: INFO nova.compute.manager [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Terminating instance [ 1196.912777] env[61906]: DEBUG nova.compute.manager [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1196.912975] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1196.913829] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58d01a2a-55a9-419b-b558-ac9c2b3bd109 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.916757] env[61906]: INFO nova.compute.manager [-] [instance: d4a6abdd-75ee-4040-8ca8-1863dd40695c] Took 1.45 seconds to deallocate network for instance. [ 1196.923481] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1196.923867] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d1432001-b74c-45d7-83f2-da5c127423af {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.929842] env[61906]: DEBUG oslo_vmware.api [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1196.929842] env[61906]: value = "task-1333831" [ 1196.929842] env[61906]: _type = "Task" [ 1196.929842] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.937088] env[61906]: DEBUG oslo_vmware.api [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333831, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.426049] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.426409] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.426528] env[61906]: DEBUG nova.objects.instance [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lazy-loading 'resources' on Instance uuid d4a6abdd-75ee-4040-8ca8-1863dd40695c {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1197.439647] env[61906]: DEBUG oslo_vmware.api [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333831, 'name': PowerOffVM_Task, 'duration_secs': 0.193517} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.439886] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1197.440072] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1197.440308] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-21d49ced-377d-4eb6-9221-d0499006180e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.498604] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1197.498957] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1197.499226] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Deleting the datastore file [datastore2] e75681a9-c315-4f60-9095-fa60e7e39c22 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1197.499501] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-32482bee-fa08-48e5-a722-dbaba46259cd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.508303] env[61906]: DEBUG oslo_vmware.api [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1197.508303] env[61906]: value = "task-1333833" [ 1197.508303] env[61906]: _type = "Task" [ 1197.508303] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.517121] env[61906]: DEBUG oslo_vmware.api [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333833, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.995569] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c877ccd-3f36-45cb-9d95-bf74b02ec3ed {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.003012] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76a11600-eae9-417b-add3-b906feacb41a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.034545] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-452612c9-6293-4454-a890-48b5d876d9db {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.043877] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bce8b4e-0a34-456c-826b-afce1ef9975d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.047374] env[61906]: DEBUG oslo_vmware.api [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333833, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145781} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.047621] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1198.047836] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1198.048040] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1198.048226] env[61906]: INFO nova.compute.manager [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1198.048459] env[61906]: DEBUG oslo.service.loopingcall [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1198.048941] env[61906]: DEBUG nova.compute.manager [-] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1198.049055] env[61906]: DEBUG nova.network.neutron [-] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1198.057789] env[61906]: DEBUG nova.compute.provider_tree [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1198.475117] env[61906]: DEBUG nova.compute.manager [req-ebc30574-7566-4f23-8f62-23b3480bb063 req-9284be30-b390-4251-84b2-bcfe87bece9d service nova] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Received event network-vif-deleted-1b29cb2d-b70e-4022-b7c7-fd014b216c1c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1198.475362] env[61906]: INFO nova.compute.manager [req-ebc30574-7566-4f23-8f62-23b3480bb063 req-9284be30-b390-4251-84b2-bcfe87bece9d service nova] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Neutron deleted interface 1b29cb2d-b70e-4022-b7c7-fd014b216c1c; detaching it from the instance and deleting it from the info cache [ 1198.475549] env[61906]: DEBUG nova.network.neutron [req-ebc30574-7566-4f23-8f62-23b3480bb063 req-9284be30-b390-4251-84b2-bcfe87bece9d service nova] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1198.560772] env[61906]: DEBUG nova.scheduler.client.report [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1198.955881] env[61906]: DEBUG nova.network.neutron [-] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1198.977441] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b00b794b-9c8f-41a6-8889-88c90880e5ac {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.987961] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73381523-971a-43d3-95d2-e189cb177333 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.012813] env[61906]: DEBUG nova.compute.manager [req-ebc30574-7566-4f23-8f62-23b3480bb063 req-9284be30-b390-4251-84b2-bcfe87bece9d service nova] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Detach interface failed, port_id=1b29cb2d-b70e-4022-b7c7-fd014b216c1c, reason: Instance e75681a9-c315-4f60-9095-fa60e7e39c22 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1199.067717] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.641s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1199.088795] env[61906]: INFO nova.scheduler.client.report [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Deleted allocations for instance d4a6abdd-75ee-4040-8ca8-1863dd40695c [ 1199.460315] env[61906]: INFO nova.compute.manager [-] [instance: e75681a9-c315-4f60-9095-fa60e7e39c22] Took 1.41 seconds to deallocate network for instance. [ 1199.599723] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7ddd79af-9960-49a4-8b66-28f6145122b4 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "d4a6abdd-75ee-4040-8ca8-1863dd40695c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.242s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1199.799462] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9d2012c2-a6c1-4da6-869f-0af40eb9aad3 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1199.799753] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9d2012c2-a6c1-4da6-869f-0af40eb9aad3 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1199.966246] env[61906]: DEBUG oslo_concurrency.lockutils [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1199.966474] env[61906]: DEBUG oslo_concurrency.lockutils [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1199.966680] env[61906]: DEBUG oslo_concurrency.lockutils [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1199.987345] env[61906]: INFO nova.scheduler.client.report [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Deleted allocations for instance e75681a9-c315-4f60-9095-fa60e7e39c22 [ 1200.304444] env[61906]: INFO nova.compute.manager [None req-9d2012c2-a6c1-4da6-869f-0af40eb9aad3 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Detaching volume 3e22fc47-57b1-4c7b-b334-e0e05317199f [ 1200.334534] env[61906]: INFO nova.virt.block_device [None req-9d2012c2-a6c1-4da6-869f-0af40eb9aad3 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Attempting to driver detach volume 3e22fc47-57b1-4c7b-b334-e0e05317199f from mountpoint /dev/sdb [ 1200.334825] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d2012c2-a6c1-4da6-869f-0af40eb9aad3 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Volume detach. Driver type: vmdk {{(pid=61906) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1200.334977] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d2012c2-a6c1-4da6-869f-0af40eb9aad3 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284850', 'volume_id': '3e22fc47-57b1-4c7b-b334-e0e05317199f', 'name': 'volume-3e22fc47-57b1-4c7b-b334-e0e05317199f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '151cc6f4-96c5-4ac3-97d5-73cddfa7b465', 'attached_at': '', 'detached_at': '', 'volume_id': '3e22fc47-57b1-4c7b-b334-e0e05317199f', 'serial': '3e22fc47-57b1-4c7b-b334-e0e05317199f'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1200.335915] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca74b505-5b32-421c-af56-78f25840ecc1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.358435] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c8961ea-89ad-49c2-a79d-3befdc572202 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.365018] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-742566a9-24c0-41f0-aac0-62fca1928094 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.384584] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59ebad91-4c85-4cac-8628-59456b9dfe65 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.399069] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d2012c2-a6c1-4da6-869f-0af40eb9aad3 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] The volume has not been displaced from its original location: [datastore1] volume-3e22fc47-57b1-4c7b-b334-e0e05317199f/volume-3e22fc47-57b1-4c7b-b334-e0e05317199f.vmdk. No consolidation needed. {{(pid=61906) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1200.404270] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d2012c2-a6c1-4da6-869f-0af40eb9aad3 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Reconfiguring VM instance instance-00000068 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1200.404549] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f794135f-f631-4265-9fdf-1618e939445c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.421472] env[61906]: DEBUG oslo_vmware.api [None req-9d2012c2-a6c1-4da6-869f-0af40eb9aad3 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1200.421472] env[61906]: value = "task-1333835" [ 1200.421472] env[61906]: _type = "Task" [ 1200.421472] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.428677] env[61906]: DEBUG oslo_vmware.api [None req-9d2012c2-a6c1-4da6-869f-0af40eb9aad3 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333835, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.495317] env[61906]: DEBUG oslo_concurrency.lockutils [None req-21154e01-b670-426a-b589-31cb0eb2496d tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "e75681a9-c315-4f60-9095-fa60e7e39c22" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.587s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1200.931682] env[61906]: DEBUG oslo_vmware.api [None req-9d2012c2-a6c1-4da6-869f-0af40eb9aad3 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333835, 'name': ReconfigVM_Task, 'duration_secs': 0.198425} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.931963] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d2012c2-a6c1-4da6-869f-0af40eb9aad3 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Reconfigured VM instance instance-00000068 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1200.936535] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-befd61d1-11f3-4e1a-898e-1384fedc7884 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.951023] env[61906]: DEBUG oslo_vmware.api [None req-9d2012c2-a6c1-4da6-869f-0af40eb9aad3 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1200.951023] env[61906]: value = "task-1333836" [ 1200.951023] env[61906]: _type = "Task" [ 1200.951023] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.958423] env[61906]: DEBUG oslo_vmware.api [None req-9d2012c2-a6c1-4da6-869f-0af40eb9aad3 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333836, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.460778] env[61906]: DEBUG oslo_vmware.api [None req-9d2012c2-a6c1-4da6-869f-0af40eb9aad3 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333836, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.961257] env[61906]: DEBUG oslo_vmware.api [None req-9d2012c2-a6c1-4da6-869f-0af40eb9aad3 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333836, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.997265] env[61906]: DEBUG oslo_concurrency.lockutils [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1201.997496] env[61906]: DEBUG oslo_concurrency.lockutils [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1202.463601] env[61906]: DEBUG oslo_vmware.api [None req-9d2012c2-a6c1-4da6-869f-0af40eb9aad3 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333836, 'name': ReconfigVM_Task, 'duration_secs': 1.129255} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.463924] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d2012c2-a6c1-4da6-869f-0af40eb9aad3 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284850', 'volume_id': '3e22fc47-57b1-4c7b-b334-e0e05317199f', 'name': 'volume-3e22fc47-57b1-4c7b-b334-e0e05317199f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '151cc6f4-96c5-4ac3-97d5-73cddfa7b465', 'attached_at': '', 'detached_at': '', 'volume_id': '3e22fc47-57b1-4c7b-b334-e0e05317199f', 'serial': '3e22fc47-57b1-4c7b-b334-e0e05317199f'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1202.499418] env[61906]: DEBUG nova.compute.manager [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1203.006860] env[61906]: DEBUG nova.objects.instance [None req-9d2012c2-a6c1-4da6-869f-0af40eb9aad3 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lazy-loading 'flavor' on Instance uuid 151cc6f4-96c5-4ac3-97d5-73cddfa7b465 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1203.023571] env[61906]: DEBUG oslo_concurrency.lockutils [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.023836] env[61906]: DEBUG oslo_concurrency.lockutils [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.025392] env[61906]: INFO nova.compute.claims [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1204.016102] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9d2012c2-a6c1-4da6-869f-0af40eb9aad3 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.216s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1204.086342] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3e1575b-face-461c-a574-bc6349e57abb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.094445] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc896ded-5be8-43d8-82e7-4b92a75da68e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.123549] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fca4221-95a4-4e68-8bc8-61595db13a8d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.130090] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-392c0d6c-6392-415f-b4fb-9d1524ecb142 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.142433] env[61906]: DEBUG nova.compute.provider_tree [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1204.328343] env[61906]: DEBUG oslo_concurrency.lockutils [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1204.328706] env[61906]: DEBUG oslo_concurrency.lockutils [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1204.328814] env[61906]: DEBUG oslo_concurrency.lockutils [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1204.328987] env[61906]: DEBUG oslo_concurrency.lockutils [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1204.329177] env[61906]: DEBUG oslo_concurrency.lockutils [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1204.331164] env[61906]: INFO nova.compute.manager [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Terminating instance [ 1204.332834] env[61906]: DEBUG nova.compute.manager [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1204.333039] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1204.333879] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-357768a3-a297-435d-a56e-7d012c9b88a0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.341416] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1204.341872] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-52ec931b-370d-49ef-8708-e314f26e1e19 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.347358] env[61906]: DEBUG oslo_vmware.api [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1204.347358] env[61906]: value = "task-1333837" [ 1204.347358] env[61906]: _type = "Task" [ 1204.347358] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.354855] env[61906]: DEBUG oslo_vmware.api [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333837, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.645541] env[61906]: DEBUG nova.scheduler.client.report [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1204.856956] env[61906]: DEBUG oslo_vmware.api [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333837, 'name': PowerOffVM_Task, 'duration_secs': 0.224471} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.857225] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1204.857403] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1204.857648] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-992710be-1ccc-48d2-b2f9-881b5c91af22 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.915701] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1204.915925] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1204.916203] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Deleting the datastore file [datastore2] f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1204.916455] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8df6556e-bd40-444c-8759-babfa6084173 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.922208] env[61906]: DEBUG oslo_vmware.api [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1204.922208] env[61906]: value = "task-1333839" [ 1204.922208] env[61906]: _type = "Task" [ 1204.922208] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.929547] env[61906]: DEBUG oslo_vmware.api [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333839, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.061606] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1205.061922] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1205.063036] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1205.063036] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1205.063036] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1205.064649] env[61906]: INFO nova.compute.manager [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Terminating instance [ 1205.066487] env[61906]: DEBUG nova.compute.manager [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1205.066704] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1205.067549] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-854652d7-98d2-42c1-bc53-f5ac4a777ff2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.075019] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1205.075272] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3b21ff39-341e-4361-a867-2ebcf0215cec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.082492] env[61906]: DEBUG oslo_vmware.api [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1205.082492] env[61906]: value = "task-1333840" [ 1205.082492] env[61906]: _type = "Task" [ 1205.082492] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.090035] env[61906]: DEBUG oslo_vmware.api [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333840, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.150161] env[61906]: DEBUG oslo_concurrency.lockutils [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.126s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1205.150653] env[61906]: DEBUG nova.compute.manager [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1205.432403] env[61906]: DEBUG oslo_vmware.api [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333839, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.183129} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.435182] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1205.435182] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1205.435182] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1205.435182] env[61906]: INFO nova.compute.manager [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1205.435182] env[61906]: DEBUG oslo.service.loopingcall [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1205.435182] env[61906]: DEBUG nova.compute.manager [-] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1205.435182] env[61906]: DEBUG nova.network.neutron [-] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1205.594355] env[61906]: DEBUG oslo_vmware.api [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333840, 'name': PowerOffVM_Task, 'duration_secs': 0.193676} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.595034] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1205.595034] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1205.595034] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-de38d53f-ee2e-4f8d-add2-d5f3c3dcbb44 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.657269] env[61906]: DEBUG nova.compute.utils [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1205.660365] env[61906]: DEBUG nova.compute.manager [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1205.660596] env[61906]: DEBUG nova.network.neutron [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1205.662823] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1205.663059] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1205.663321] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Deleting the datastore file [datastore1] 151cc6f4-96c5-4ac3-97d5-73cddfa7b465 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1205.664066] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3eeaee71-6204-419a-9fa4-6beed1a8b45c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.670059] env[61906]: DEBUG nova.compute.manager [req-9da443c4-721b-4f22-a470-25ebd6aa86a2 req-611bcc7a-cb09-4727-8be0-5cbab1bdf0be service nova] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Received event network-vif-deleted-697dbd16-ab06-44ba-ab39-8da5e1ce0c78 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1205.670298] env[61906]: INFO nova.compute.manager [req-9da443c4-721b-4f22-a470-25ebd6aa86a2 req-611bcc7a-cb09-4727-8be0-5cbab1bdf0be service nova] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Neutron deleted interface 697dbd16-ab06-44ba-ab39-8da5e1ce0c78; detaching it from the instance and deleting it from the info cache [ 1205.670542] env[61906]: DEBUG nova.network.neutron [req-9da443c4-721b-4f22-a470-25ebd6aa86a2 req-611bcc7a-cb09-4727-8be0-5cbab1bdf0be service nova] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1205.673841] env[61906]: DEBUG oslo_vmware.api [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1205.673841] env[61906]: value = "task-1333842" [ 1205.673841] env[61906]: _type = "Task" [ 1205.673841] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.685128] env[61906]: DEBUG oslo_vmware.api [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333842, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.711464] env[61906]: DEBUG nova.policy [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '27ba0d3b00ce4513a03259970a600190', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b7d3aa88849f4492aa16ff337b8cfdda', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1206.000251] env[61906]: DEBUG nova.network.neutron [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Successfully created port: dba2c370-204e-45ba-9ce0-793505bd03f5 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1206.147968] env[61906]: DEBUG nova.network.neutron [-] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1206.164501] env[61906]: DEBUG nova.compute.manager [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1206.175181] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4b22f35a-9939-4558-8bc9-c9345d1d6784 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.186700] env[61906]: DEBUG oslo_vmware.api [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333842, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131027} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.187786] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1206.187982] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1206.188182] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1206.188358] env[61906]: INFO nova.compute.manager [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1206.188602] env[61906]: DEBUG oslo.service.loopingcall [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1206.188849] env[61906]: DEBUG nova.compute.manager [-] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1206.188941] env[61906]: DEBUG nova.network.neutron [-] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1206.192935] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2ecdb06-0a37-4287-ad90-99760212dc29 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.215628] env[61906]: DEBUG nova.compute.manager [req-9da443c4-721b-4f22-a470-25ebd6aa86a2 req-611bcc7a-cb09-4727-8be0-5cbab1bdf0be service nova] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Detach interface failed, port_id=697dbd16-ab06-44ba-ab39-8da5e1ce0c78, reason: Instance f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1206.651340] env[61906]: INFO nova.compute.manager [-] [instance: f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b] Took 1.22 seconds to deallocate network for instance. [ 1206.660772] env[61906]: DEBUG nova.compute.manager [req-ff10c869-1687-4ef0-97be-c646e948e31c req-49036612-61e6-497a-9cf3-19d897f07d54 service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Received event network-vif-deleted-1d0e39a6-9bf4-4034-b7e3-0d500971fc09 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1206.660772] env[61906]: INFO nova.compute.manager [req-ff10c869-1687-4ef0-97be-c646e948e31c req-49036612-61e6-497a-9cf3-19d897f07d54 service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Neutron deleted interface 1d0e39a6-9bf4-4034-b7e3-0d500971fc09; detaching it from the instance and deleting it from the info cache [ 1206.660901] env[61906]: DEBUG nova.network.neutron [req-ff10c869-1687-4ef0-97be-c646e948e31c req-49036612-61e6-497a-9cf3-19d897f07d54 service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1207.107282] env[61906]: DEBUG nova.network.neutron [-] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1207.159151] env[61906]: DEBUG oslo_concurrency.lockutils [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1207.159151] env[61906]: DEBUG oslo_concurrency.lockutils [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1207.159151] env[61906]: DEBUG nova.objects.instance [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lazy-loading 'resources' on Instance uuid f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1207.163524] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bb92df1f-1a2b-4cfe-bd29-cfceaa478ecc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.175537] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19d7416d-c42b-4b1f-afc5-d791151d7141 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.186655] env[61906]: DEBUG nova.compute.manager [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1207.202173] env[61906]: DEBUG nova.compute.manager [req-ff10c869-1687-4ef0-97be-c646e948e31c req-49036612-61e6-497a-9cf3-19d897f07d54 service nova] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Detach interface failed, port_id=1d0e39a6-9bf4-4034-b7e3-0d500971fc09, reason: Instance 151cc6f4-96c5-4ac3-97d5-73cddfa7b465 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1207.213035] env[61906]: DEBUG nova.virt.hardware [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1207.213035] env[61906]: DEBUG nova.virt.hardware [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1207.213035] env[61906]: DEBUG nova.virt.hardware [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1207.213363] env[61906]: DEBUG nova.virt.hardware [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1207.213363] env[61906]: DEBUG nova.virt.hardware [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1207.213483] env[61906]: DEBUG nova.virt.hardware [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1207.213743] env[61906]: DEBUG nova.virt.hardware [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1207.214379] env[61906]: DEBUG nova.virt.hardware [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1207.214629] env[61906]: DEBUG nova.virt.hardware [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1207.214920] env[61906]: DEBUG nova.virt.hardware [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1207.215153] env[61906]: DEBUG nova.virt.hardware [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1207.216045] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eebf0ff5-855d-420e-9231-1ebfd9dd5557 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.224143] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff7e21c-1673-4d55-a4be-73da1494d4b9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.456670] env[61906]: DEBUG nova.network.neutron [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Successfully updated port: dba2c370-204e-45ba-9ce0-793505bd03f5 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1207.611490] env[61906]: INFO nova.compute.manager [-] [instance: 151cc6f4-96c5-4ac3-97d5-73cddfa7b465] Took 1.42 seconds to deallocate network for instance. [ 1207.699301] env[61906]: DEBUG nova.compute.manager [req-df2956e9-2941-4a86-b72a-1b6a5cf0814e req-f125b821-443f-4993-93bf-eff0c0ad043b service nova] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Received event network-vif-plugged-dba2c370-204e-45ba-9ce0-793505bd03f5 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1207.699597] env[61906]: DEBUG oslo_concurrency.lockutils [req-df2956e9-2941-4a86-b72a-1b6a5cf0814e req-f125b821-443f-4993-93bf-eff0c0ad043b service nova] Acquiring lock "8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1207.699817] env[61906]: DEBUG oslo_concurrency.lockutils [req-df2956e9-2941-4a86-b72a-1b6a5cf0814e req-f125b821-443f-4993-93bf-eff0c0ad043b service nova] Lock "8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1207.700029] env[61906]: DEBUG oslo_concurrency.lockutils [req-df2956e9-2941-4a86-b72a-1b6a5cf0814e req-f125b821-443f-4993-93bf-eff0c0ad043b service nova] Lock "8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1207.700179] env[61906]: DEBUG nova.compute.manager [req-df2956e9-2941-4a86-b72a-1b6a5cf0814e req-f125b821-443f-4993-93bf-eff0c0ad043b service nova] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] No waiting events found dispatching network-vif-plugged-dba2c370-204e-45ba-9ce0-793505bd03f5 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1207.700348] env[61906]: WARNING nova.compute.manager [req-df2956e9-2941-4a86-b72a-1b6a5cf0814e req-f125b821-443f-4993-93bf-eff0c0ad043b service nova] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Received unexpected event network-vif-plugged-dba2c370-204e-45ba-9ce0-793505bd03f5 for instance with vm_state building and task_state spawning. [ 1207.700512] env[61906]: DEBUG nova.compute.manager [req-df2956e9-2941-4a86-b72a-1b6a5cf0814e req-f125b821-443f-4993-93bf-eff0c0ad043b service nova] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Received event network-changed-dba2c370-204e-45ba-9ce0-793505bd03f5 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1207.700695] env[61906]: DEBUG nova.compute.manager [req-df2956e9-2941-4a86-b72a-1b6a5cf0814e req-f125b821-443f-4993-93bf-eff0c0ad043b service nova] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Refreshing instance network info cache due to event network-changed-dba2c370-204e-45ba-9ce0-793505bd03f5. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1207.700900] env[61906]: DEBUG oslo_concurrency.lockutils [req-df2956e9-2941-4a86-b72a-1b6a5cf0814e req-f125b821-443f-4993-93bf-eff0c0ad043b service nova] Acquiring lock "refresh_cache-8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1207.701072] env[61906]: DEBUG oslo_concurrency.lockutils [req-df2956e9-2941-4a86-b72a-1b6a5cf0814e req-f125b821-443f-4993-93bf-eff0c0ad043b service nova] Acquired lock "refresh_cache-8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1207.701250] env[61906]: DEBUG nova.network.neutron [req-df2956e9-2941-4a86-b72a-1b6a5cf0814e req-f125b821-443f-4993-93bf-eff0c0ad043b service nova] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Refreshing network info cache for port dba2c370-204e-45ba-9ce0-793505bd03f5 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1207.730286] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce6b15fd-5fa2-48d1-9e5a-466a8dc01f42 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.738674] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f60be786-6358-443b-a38f-ec2d864254da {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.769006] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd30ea22-a932-4579-936d-25b04b34be79 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.777312] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a110f9c9-dcd3-4022-8790-28863f1907d6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.790403] env[61906]: DEBUG nova.compute.provider_tree [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1207.959533] env[61906]: DEBUG oslo_concurrency.lockutils [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "refresh_cache-8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1208.116078] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1208.234168] env[61906]: DEBUG nova.network.neutron [req-df2956e9-2941-4a86-b72a-1b6a5cf0814e req-f125b821-443f-4993-93bf-eff0c0ad043b service nova] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1208.293733] env[61906]: DEBUG nova.scheduler.client.report [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1208.318112] env[61906]: DEBUG nova.network.neutron [req-df2956e9-2941-4a86-b72a-1b6a5cf0814e req-f125b821-443f-4993-93bf-eff0c0ad043b service nova] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1208.798177] env[61906]: DEBUG oslo_concurrency.lockutils [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.639s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1208.800480] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.685s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1208.800723] env[61906]: DEBUG nova.objects.instance [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lazy-loading 'resources' on Instance uuid 151cc6f4-96c5-4ac3-97d5-73cddfa7b465 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1208.820325] env[61906]: INFO nova.scheduler.client.report [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Deleted allocations for instance f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b [ 1208.821462] env[61906]: DEBUG oslo_concurrency.lockutils [req-df2956e9-2941-4a86-b72a-1b6a5cf0814e req-f125b821-443f-4993-93bf-eff0c0ad043b service nova] Releasing lock "refresh_cache-8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1208.824085] env[61906]: DEBUG oslo_concurrency.lockutils [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired lock "refresh_cache-8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1208.824085] env[61906]: DEBUG nova.network.neutron [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1209.331573] env[61906]: DEBUG oslo_concurrency.lockutils [None req-68640057-3155-4218-9fae-0b7d305b9549 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "f1c2c59c-df5f-4e4a-a6c6-1251f95aed4b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.003s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1209.352775] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06929ead-9109-4386-9687-f62bd282b5de {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.357175] env[61906]: DEBUG nova.network.neutron [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1209.361909] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30f48ba6-dc13-4969-bb5d-8ec60b817cde {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.400616] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddf09458-8af2-4602-8b54-789d7a82ad3f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.409157] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7e6f767-ef5b-4c13-82ff-e8c252467112 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.422805] env[61906]: DEBUG nova.compute.provider_tree [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1209.515447] env[61906]: DEBUG nova.network.neutron [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Updating instance_info_cache with network_info: [{"id": "dba2c370-204e-45ba-9ce0-793505bd03f5", "address": "fa:16:3e:a9:97:42", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdba2c370-20", "ovs_interfaceid": "dba2c370-204e-45ba-9ce0-793505bd03f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1209.869765] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "86be7197-c0d1-49c5-aa30-cf908a506031" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1209.870143] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "86be7197-c0d1-49c5-aa30-cf908a506031" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1209.870408] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "86be7197-c0d1-49c5-aa30-cf908a506031-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1209.870637] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "86be7197-c0d1-49c5-aa30-cf908a506031-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1209.870821] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "86be7197-c0d1-49c5-aa30-cf908a506031-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1209.873733] env[61906]: INFO nova.compute.manager [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Terminating instance [ 1209.875472] env[61906]: DEBUG nova.compute.manager [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1209.875700] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1209.876570] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de5565f-259f-49e9-9eff-3d003289ac31 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.884104] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1209.884333] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7935521e-5cf9-4e23-a255-1a59843790e8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.890218] env[61906]: DEBUG oslo_vmware.api [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1209.890218] env[61906]: value = "task-1333843" [ 1209.890218] env[61906]: _type = "Task" [ 1209.890218] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.897189] env[61906]: DEBUG oslo_vmware.api [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333843, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.927773] env[61906]: DEBUG nova.scheduler.client.report [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1210.020587] env[61906]: DEBUG oslo_concurrency.lockutils [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Releasing lock "refresh_cache-8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1210.021097] env[61906]: DEBUG nova.compute.manager [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Instance network_info: |[{"id": "dba2c370-204e-45ba-9ce0-793505bd03f5", "address": "fa:16:3e:a9:97:42", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdba2c370-20", "ovs_interfaceid": "dba2c370-204e-45ba-9ce0-793505bd03f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1210.021590] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a9:97:42', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dba2c370-204e-45ba-9ce0-793505bd03f5', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1210.030872] env[61906]: DEBUG oslo.service.loopingcall [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1210.031148] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1210.031381] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8c49fa5b-3ea0-4a8d-8969-c8ef74928e08 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.053553] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1210.053553] env[61906]: value = "task-1333844" [ 1210.053553] env[61906]: _type = "Task" [ 1210.053553] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.061783] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333844, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.400512] env[61906]: DEBUG oslo_vmware.api [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333843, 'name': PowerOffVM_Task, 'duration_secs': 0.197743} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.400915] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1210.400964] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1210.401251] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d2c2e12d-1da0-40be-92da-668f20ff9d43 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.433291] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.633s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1210.457609] env[61906]: INFO nova.scheduler.client.report [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Deleted allocations for instance 151cc6f4-96c5-4ac3-97d5-73cddfa7b465 [ 1210.459388] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1210.459388] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1210.459388] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Deleting the datastore file [datastore2] 86be7197-c0d1-49c5-aa30-cf908a506031 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1210.462041] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2e739a82-062d-43ed-a610-c4e6ad084960 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.468724] env[61906]: DEBUG oslo_vmware.api [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for the task: (returnval){ [ 1210.468724] env[61906]: value = "task-1333846" [ 1210.468724] env[61906]: _type = "Task" [ 1210.468724] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.476232] env[61906]: DEBUG oslo_vmware.api [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333846, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.564048] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333844, 'name': CreateVM_Task, 'duration_secs': 0.337584} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.564229] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1210.564914] env[61906]: DEBUG oslo_concurrency.lockutils [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1210.565099] env[61906]: DEBUG oslo_concurrency.lockutils [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1210.565449] env[61906]: DEBUG oslo_concurrency.lockutils [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1210.565724] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b0e864c-18dd-4733-8c5f-2de9db379431 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.569935] env[61906]: DEBUG oslo_vmware.api [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1210.569935] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52dbf6c7-3ec9-0bbc-8a64-f8551625c29c" [ 1210.569935] env[61906]: _type = "Task" [ 1210.569935] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.577353] env[61906]: DEBUG oslo_vmware.api [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52dbf6c7-3ec9-0bbc-8a64-f8551625c29c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.967557] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e03639ca-38ba-43fa-8118-5dc6e8c33994 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "151cc6f4-96c5-4ac3-97d5-73cddfa7b465" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.906s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1210.977560] env[61906]: DEBUG oslo_vmware.api [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Task: {'id': task-1333846, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138899} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.977831] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1210.978030] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1210.978211] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1210.978385] env[61906]: INFO nova.compute.manager [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1210.978627] env[61906]: DEBUG oslo.service.loopingcall [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1210.978822] env[61906]: DEBUG nova.compute.manager [-] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1210.978918] env[61906]: DEBUG nova.network.neutron [-] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1211.081252] env[61906]: DEBUG oslo_vmware.api [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52dbf6c7-3ec9-0bbc-8a64-f8551625c29c, 'name': SearchDatastore_Task, 'duration_secs': 0.008962} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.081543] env[61906]: DEBUG oslo_concurrency.lockutils [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1211.081777] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1211.082025] env[61906]: DEBUG oslo_concurrency.lockutils [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1211.082183] env[61906]: DEBUG oslo_concurrency.lockutils [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1211.082369] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1211.082632] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-345c15c4-9a37-4e58-adfa-07e99caefa28 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.090939] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1211.091140] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1211.091860] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cee11d7d-3756-4323-b56a-06238539c13b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.097598] env[61906]: DEBUG oslo_vmware.api [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1211.097598] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52f3288d-6de1-8b06-d95b-63db0aa1d7a9" [ 1211.097598] env[61906]: _type = "Task" [ 1211.097598] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.104943] env[61906]: DEBUG oslo_vmware.api [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52f3288d-6de1-8b06-d95b-63db0aa1d7a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.208145] env[61906]: DEBUG nova.compute.manager [req-59c8c758-0b7f-463e-84ea-3e1134a39408 req-956c73a5-4374-4f9f-8fed-a3ec46958206 service nova] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Received event network-vif-deleted-24ab450e-e9c4-4d9a-9c4d-13dee30eb56b {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1211.208145] env[61906]: INFO nova.compute.manager [req-59c8c758-0b7f-463e-84ea-3e1134a39408 req-956c73a5-4374-4f9f-8fed-a3ec46958206 service nova] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Neutron deleted interface 24ab450e-e9c4-4d9a-9c4d-13dee30eb56b; detaching it from the instance and deleting it from the info cache [ 1211.208145] env[61906]: DEBUG nova.network.neutron [req-59c8c758-0b7f-463e-84ea-3e1134a39408 req-956c73a5-4374-4f9f-8fed-a3ec46958206 service nova] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1211.608801] env[61906]: DEBUG oslo_vmware.api [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52f3288d-6de1-8b06-d95b-63db0aa1d7a9, 'name': SearchDatastore_Task, 'duration_secs': 0.012174} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.609669] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4adf3bae-405a-43b3-b84e-609fae038f53 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.614672] env[61906]: DEBUG oslo_vmware.api [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1211.614672] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5289a69b-79a0-a329-b171-a5cd25fe7409" [ 1211.614672] env[61906]: _type = "Task" [ 1211.614672] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.622916] env[61906]: DEBUG oslo_vmware.api [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5289a69b-79a0-a329-b171-a5cd25fe7409, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.685541] env[61906]: DEBUG nova.network.neutron [-] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1211.710853] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7793f432-2831-4559-82b6-4f749652d5e3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.721018] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-414aef7e-3edd-4e61-98eb-90b975fe359d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.743756] env[61906]: DEBUG nova.compute.manager [req-59c8c758-0b7f-463e-84ea-3e1134a39408 req-956c73a5-4374-4f9f-8fed-a3ec46958206 service nova] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Detach interface failed, port_id=24ab450e-e9c4-4d9a-9c4d-13dee30eb56b, reason: Instance 86be7197-c0d1-49c5-aa30-cf908a506031 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1212.126862] env[61906]: DEBUG oslo_vmware.api [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]5289a69b-79a0-a329-b171-a5cd25fe7409, 'name': SearchDatastore_Task, 'duration_secs': 0.009191} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.127162] env[61906]: DEBUG oslo_concurrency.lockutils [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1212.127424] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd/8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1212.127687] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-72a6f7d5-2c29-4650-a2cd-af763aca6a1e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.134781] env[61906]: DEBUG oslo_vmware.api [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1212.134781] env[61906]: value = "task-1333847" [ 1212.134781] env[61906]: _type = "Task" [ 1212.134781] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.142411] env[61906]: DEBUG oslo_vmware.api [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333847, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.188122] env[61906]: INFO nova.compute.manager [-] [instance: 86be7197-c0d1-49c5-aa30-cf908a506031] Took 1.21 seconds to deallocate network for instance. [ 1212.644685] env[61906]: DEBUG oslo_vmware.api [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333847, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.453585} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.645044] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore1] 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd/8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1212.645201] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1212.645437] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fb1bc585-b7a9-4e87-bb63-4e5d6f970b42 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.652017] env[61906]: DEBUG oslo_vmware.api [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1212.652017] env[61906]: value = "task-1333848" [ 1212.652017] env[61906]: _type = "Task" [ 1212.652017] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.659235] env[61906]: DEBUG oslo_vmware.api [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333848, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.695551] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1212.695887] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1212.696145] env[61906]: DEBUG nova.objects.instance [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lazy-loading 'resources' on Instance uuid 86be7197-c0d1-49c5-aa30-cf908a506031 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1213.012663] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "9f6264eb-2632-4210-b5e8-a58c875a3a85" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1213.012909] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "9f6264eb-2632-4210-b5e8-a58c875a3a85" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1213.161957] env[61906]: DEBUG oslo_vmware.api [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333848, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060949} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.162252] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1213.163013] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b2e321e-251c-4d97-a3f9-e74d573c5b4a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.184340] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd/8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1213.184570] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1a61d14f-9a04-41d6-84bb-7b518268a68d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.207185] env[61906]: DEBUG oslo_vmware.api [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1213.207185] env[61906]: value = "task-1333849" [ 1213.207185] env[61906]: _type = "Task" [ 1213.207185] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.214779] env[61906]: DEBUG oslo_vmware.api [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333849, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.249027] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-737c0c8c-8ce2-4550-88a2-93f4c71cabde {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.254455] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d319c9d-ea5d-408d-93ec-ab66e3afe743 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.285751] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f82aa66-59d2-4a7c-8931-ce116b2dfb50 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.292850] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a170771c-928f-4105-8b29-9252364bc8c9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.306071] env[61906]: DEBUG nova.compute.provider_tree [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1213.515329] env[61906]: DEBUG nova.compute.manager [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1213.717719] env[61906]: DEBUG oslo_vmware.api [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333849, 'name': ReconfigVM_Task, 'duration_secs': 0.279654} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.717719] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd/8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1213.718382] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a282e932-2bd9-4408-b1dd-6031a86e5199 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.724851] env[61906]: DEBUG oslo_vmware.api [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1213.724851] env[61906]: value = "task-1333850" [ 1213.724851] env[61906]: _type = "Task" [ 1213.724851] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.732070] env[61906]: DEBUG oslo_vmware.api [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333850, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.809828] env[61906]: DEBUG nova.scheduler.client.report [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1214.036876] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1214.234554] env[61906]: DEBUG oslo_vmware.api [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333850, 'name': Rename_Task, 'duration_secs': 0.137537} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.234833] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1214.235102] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2f760b6e-9dd6-4e16-8fd5-6247c7feb356 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.241513] env[61906]: DEBUG oslo_vmware.api [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1214.241513] env[61906]: value = "task-1333851" [ 1214.241513] env[61906]: _type = "Task" [ 1214.241513] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.248406] env[61906]: DEBUG oslo_vmware.api [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333851, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.314527] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.618s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1214.316913] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.280s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1214.318471] env[61906]: INFO nova.compute.claims [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1214.333931] env[61906]: INFO nova.scheduler.client.report [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Deleted allocations for instance 86be7197-c0d1-49c5-aa30-cf908a506031 [ 1214.752342] env[61906]: DEBUG oslo_vmware.api [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333851, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.841971] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b5356a5-281a-41e4-906a-b82f04be4977 tempest-ServerRescueNegativeTestJSON-1621303057 tempest-ServerRescueNegativeTestJSON-1621303057-project-member] Lock "86be7197-c0d1-49c5-aa30-cf908a506031" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.972s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1215.253169] env[61906]: DEBUG oslo_vmware.api [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333851, 'name': PowerOnVM_Task, 'duration_secs': 0.523704} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.254048] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1215.254048] env[61906]: INFO nova.compute.manager [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Took 8.07 seconds to spawn the instance on the hypervisor. [ 1215.254048] env[61906]: DEBUG nova.compute.manager [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1215.254584] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4fc6e7d-9eb9-4df9-95c2-e485929d5306 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.363712] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d85061-f9c3-4616-995b-5dc84ecf30a8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.371264] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d7fd9e2-f6ef-4a1b-90b1-5b229831aab1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.400621] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe86b992-3281-4c5e-a499-833c8d460c8d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.408112] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa7421eb-c910-4216-a389-e16e2e3339b2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.421789] env[61906]: DEBUG nova.compute.provider_tree [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1215.778374] env[61906]: INFO nova.compute.manager [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Took 12.77 seconds to build instance. [ 1215.926497] env[61906]: DEBUG nova.scheduler.client.report [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1216.282425] env[61906]: DEBUG oslo_concurrency.lockutils [None req-677c1eea-ccd1-473c-9f17-a7e5cec5902a tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.285s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1216.423186] env[61906]: DEBUG nova.compute.manager [req-0e65850a-58fc-4331-9aaf-0a1d87f9cef9 req-6be63841-9ee8-4a43-b7ca-4c39455b632a service nova] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Received event network-changed-dba2c370-204e-45ba-9ce0-793505bd03f5 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1216.423186] env[61906]: DEBUG nova.compute.manager [req-0e65850a-58fc-4331-9aaf-0a1d87f9cef9 req-6be63841-9ee8-4a43-b7ca-4c39455b632a service nova] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Refreshing instance network info cache due to event network-changed-dba2c370-204e-45ba-9ce0-793505bd03f5. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1216.423556] env[61906]: DEBUG oslo_concurrency.lockutils [req-0e65850a-58fc-4331-9aaf-0a1d87f9cef9 req-6be63841-9ee8-4a43-b7ca-4c39455b632a service nova] Acquiring lock "refresh_cache-8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1216.423846] env[61906]: DEBUG oslo_concurrency.lockutils [req-0e65850a-58fc-4331-9aaf-0a1d87f9cef9 req-6be63841-9ee8-4a43-b7ca-4c39455b632a service nova] Acquired lock "refresh_cache-8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1216.424150] env[61906]: DEBUG nova.network.neutron [req-0e65850a-58fc-4331-9aaf-0a1d87f9cef9 req-6be63841-9ee8-4a43-b7ca-4c39455b632a service nova] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Refreshing network info cache for port dba2c370-204e-45ba-9ce0-793505bd03f5 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1216.430610] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.114s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1216.431762] env[61906]: DEBUG nova.compute.manager [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1216.939793] env[61906]: DEBUG nova.compute.utils [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1216.941966] env[61906]: DEBUG nova.compute.manager [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1216.942622] env[61906]: DEBUG nova.network.neutron [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1216.982539] env[61906]: DEBUG nova.policy [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3896807bf4d64d08b988b6c2023a408b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '00c03ca0de1447da9d47c83f256efdcd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1217.153408] env[61906]: DEBUG nova.network.neutron [req-0e65850a-58fc-4331-9aaf-0a1d87f9cef9 req-6be63841-9ee8-4a43-b7ca-4c39455b632a service nova] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Updated VIF entry in instance network info cache for port dba2c370-204e-45ba-9ce0-793505bd03f5. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1217.153862] env[61906]: DEBUG nova.network.neutron [req-0e65850a-58fc-4331-9aaf-0a1d87f9cef9 req-6be63841-9ee8-4a43-b7ca-4c39455b632a service nova] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Updating instance_info_cache with network_info: [{"id": "dba2c370-204e-45ba-9ce0-793505bd03f5", "address": "fa:16:3e:a9:97:42", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdba2c370-20", "ovs_interfaceid": "dba2c370-204e-45ba-9ce0-793505bd03f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1217.256822] env[61906]: DEBUG nova.network.neutron [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Successfully created port: 0bc154b4-abd3-4632-8b74-c16ded1c94d0 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1217.444881] env[61906]: DEBUG nova.compute.manager [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1217.656805] env[61906]: DEBUG oslo_concurrency.lockutils [req-0e65850a-58fc-4331-9aaf-0a1d87f9cef9 req-6be63841-9ee8-4a43-b7ca-4c39455b632a service nova] Releasing lock "refresh_cache-8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1218.454918] env[61906]: DEBUG nova.compute.manager [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1218.479084] env[61906]: DEBUG nova.virt.hardware [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T06:18:03Z,direct_url=,disk_format='vmdk',id=a9502563-1f96-46a5-b87e-12c0e9375e6e,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5dc9cad1b7084726927920f91ed739c9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T06:18:04Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1218.479363] env[61906]: DEBUG nova.virt.hardware [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1218.479528] env[61906]: DEBUG nova.virt.hardware [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1218.479717] env[61906]: DEBUG nova.virt.hardware [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1218.479872] env[61906]: DEBUG nova.virt.hardware [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1218.480034] env[61906]: DEBUG nova.virt.hardware [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1218.480258] env[61906]: DEBUG nova.virt.hardware [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1218.480426] env[61906]: DEBUG nova.virt.hardware [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1218.480600] env[61906]: DEBUG nova.virt.hardware [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1218.480768] env[61906]: DEBUG nova.virt.hardware [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1218.481050] env[61906]: DEBUG nova.virt.hardware [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1218.481915] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23f3c1ba-9cdb-4336-99d8-0337470f123d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.490785] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d0c4e8e-e1c9-4657-9d52-2057581a3757 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.661270] env[61906]: DEBUG nova.compute.manager [req-4d7b0cac-4007-44c9-a14e-5d7af6747115 req-fda2f284-13ad-4c1b-b0b9-aa0865f714ec service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Received event network-vif-plugged-0bc154b4-abd3-4632-8b74-c16ded1c94d0 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1218.661503] env[61906]: DEBUG oslo_concurrency.lockutils [req-4d7b0cac-4007-44c9-a14e-5d7af6747115 req-fda2f284-13ad-4c1b-b0b9-aa0865f714ec service nova] Acquiring lock "9f6264eb-2632-4210-b5e8-a58c875a3a85-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1218.661715] env[61906]: DEBUG oslo_concurrency.lockutils [req-4d7b0cac-4007-44c9-a14e-5d7af6747115 req-fda2f284-13ad-4c1b-b0b9-aa0865f714ec service nova] Lock "9f6264eb-2632-4210-b5e8-a58c875a3a85-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1218.661892] env[61906]: DEBUG oslo_concurrency.lockutils [req-4d7b0cac-4007-44c9-a14e-5d7af6747115 req-fda2f284-13ad-4c1b-b0b9-aa0865f714ec service nova] Lock "9f6264eb-2632-4210-b5e8-a58c875a3a85-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.662073] env[61906]: DEBUG nova.compute.manager [req-4d7b0cac-4007-44c9-a14e-5d7af6747115 req-fda2f284-13ad-4c1b-b0b9-aa0865f714ec service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] No waiting events found dispatching network-vif-plugged-0bc154b4-abd3-4632-8b74-c16ded1c94d0 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1218.662247] env[61906]: WARNING nova.compute.manager [req-4d7b0cac-4007-44c9-a14e-5d7af6747115 req-fda2f284-13ad-4c1b-b0b9-aa0865f714ec service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Received unexpected event network-vif-plugged-0bc154b4-abd3-4632-8b74-c16ded1c94d0 for instance with vm_state building and task_state spawning. [ 1218.741877] env[61906]: DEBUG nova.network.neutron [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Successfully updated port: 0bc154b4-abd3-4632-8b74-c16ded1c94d0 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1219.244492] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "refresh_cache-9f6264eb-2632-4210-b5e8-a58c875a3a85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1219.244644] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquired lock "refresh_cache-9f6264eb-2632-4210-b5e8-a58c875a3a85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1219.244800] env[61906]: DEBUG nova.network.neutron [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1219.782058] env[61906]: DEBUG nova.network.neutron [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1219.916972] env[61906]: DEBUG nova.network.neutron [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Updating instance_info_cache with network_info: [{"id": "0bc154b4-abd3-4632-8b74-c16ded1c94d0", "address": "fa:16:3e:2c:d0:74", "network": {"id": "0085e7d2-e26f-4b5d-a282-cf6e0f81786a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1504444833-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00c03ca0de1447da9d47c83f256efdcd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bc154b4-ab", "ovs_interfaceid": "0bc154b4-abd3-4632-8b74-c16ded1c94d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1220.420083] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Releasing lock "refresh_cache-9f6264eb-2632-4210-b5e8-a58c875a3a85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1220.420429] env[61906]: DEBUG nova.compute.manager [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Instance network_info: |[{"id": "0bc154b4-abd3-4632-8b74-c16ded1c94d0", "address": "fa:16:3e:2c:d0:74", "network": {"id": "0085e7d2-e26f-4b5d-a282-cf6e0f81786a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1504444833-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00c03ca0de1447da9d47c83f256efdcd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bc154b4-ab", "ovs_interfaceid": "0bc154b4-abd3-4632-8b74-c16ded1c94d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1220.420898] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2c:d0:74', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c6934071-bf85-4591-9c7d-55c7ea131262', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0bc154b4-abd3-4632-8b74-c16ded1c94d0', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1220.428346] env[61906]: DEBUG oslo.service.loopingcall [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1220.428557] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1220.428777] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ba2dc254-4a0a-492d-9d62-74cf1d16c698 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.448757] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1220.448757] env[61906]: value = "task-1333852" [ 1220.448757] env[61906]: _type = "Task" [ 1220.448757] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.456065] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333852, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.690516] env[61906]: DEBUG nova.compute.manager [req-eba84f42-8415-4dcf-93a8-ca82327595f7 req-70060ed4-8cc3-4761-9cb1-d339980ff8c2 service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Received event network-changed-0bc154b4-abd3-4632-8b74-c16ded1c94d0 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1220.690721] env[61906]: DEBUG nova.compute.manager [req-eba84f42-8415-4dcf-93a8-ca82327595f7 req-70060ed4-8cc3-4761-9cb1-d339980ff8c2 service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Refreshing instance network info cache due to event network-changed-0bc154b4-abd3-4632-8b74-c16ded1c94d0. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1220.690983] env[61906]: DEBUG oslo_concurrency.lockutils [req-eba84f42-8415-4dcf-93a8-ca82327595f7 req-70060ed4-8cc3-4761-9cb1-d339980ff8c2 service nova] Acquiring lock "refresh_cache-9f6264eb-2632-4210-b5e8-a58c875a3a85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1220.691157] env[61906]: DEBUG oslo_concurrency.lockutils [req-eba84f42-8415-4dcf-93a8-ca82327595f7 req-70060ed4-8cc3-4761-9cb1-d339980ff8c2 service nova] Acquired lock "refresh_cache-9f6264eb-2632-4210-b5e8-a58c875a3a85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1220.691325] env[61906]: DEBUG nova.network.neutron [req-eba84f42-8415-4dcf-93a8-ca82327595f7 req-70060ed4-8cc3-4761-9cb1-d339980ff8c2 service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Refreshing network info cache for port 0bc154b4-abd3-4632-8b74-c16ded1c94d0 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1220.959898] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333852, 'name': CreateVM_Task, 'duration_secs': 0.314462} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.960258] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1220.960691] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1220.960864] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1220.961223] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1220.961464] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-270ef5db-39a1-491a-acb1-07a836c61b40 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.966096] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1220.966096] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52da00bf-065b-40b3-ad07-59d19daf23af" [ 1220.966096] env[61906]: _type = "Task" [ 1220.966096] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.973545] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52da00bf-065b-40b3-ad07-59d19daf23af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.403449] env[61906]: DEBUG nova.network.neutron [req-eba84f42-8415-4dcf-93a8-ca82327595f7 req-70060ed4-8cc3-4761-9cb1-d339980ff8c2 service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Updated VIF entry in instance network info cache for port 0bc154b4-abd3-4632-8b74-c16ded1c94d0. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1221.403942] env[61906]: DEBUG nova.network.neutron [req-eba84f42-8415-4dcf-93a8-ca82327595f7 req-70060ed4-8cc3-4761-9cb1-d339980ff8c2 service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Updating instance_info_cache with network_info: [{"id": "0bc154b4-abd3-4632-8b74-c16ded1c94d0", "address": "fa:16:3e:2c:d0:74", "network": {"id": "0085e7d2-e26f-4b5d-a282-cf6e0f81786a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1504444833-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00c03ca0de1447da9d47c83f256efdcd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bc154b4-ab", "ovs_interfaceid": "0bc154b4-abd3-4632-8b74-c16ded1c94d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1221.477168] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52da00bf-065b-40b3-ad07-59d19daf23af, 'name': SearchDatastore_Task, 'duration_secs': 0.009792} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.477501] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1221.477744] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Processing image a9502563-1f96-46a5-b87e-12c0e9375e6e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1221.477980] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1221.478165] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1221.478359] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1221.478639] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5d7b14d0-9eea-4ebf-a6d2-dd27c9c39534 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.487831] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1221.488056] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1221.488815] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cef8f105-dc5c-47fc-90f0-8e37f0f40369 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.494022] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1221.494022] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]520c15ea-471c-5d65-7eb8-44266013a10c" [ 1221.494022] env[61906]: _type = "Task" [ 1221.494022] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.501642] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]520c15ea-471c-5d65-7eb8-44266013a10c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.906563] env[61906]: DEBUG oslo_concurrency.lockutils [req-eba84f42-8415-4dcf-93a8-ca82327595f7 req-70060ed4-8cc3-4761-9cb1-d339980ff8c2 service nova] Releasing lock "refresh_cache-9f6264eb-2632-4210-b5e8-a58c875a3a85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1222.004074] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]520c15ea-471c-5d65-7eb8-44266013a10c, 'name': SearchDatastore_Task, 'duration_secs': 0.009868} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.004842] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54e47748-9a17-4aca-bc67-5b2fa2740844 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.009806] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1222.009806] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52072911-1a6b-046a-8d0c-f7f9eaf05b6f" [ 1222.009806] env[61906]: _type = "Task" [ 1222.009806] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.018142] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52072911-1a6b-046a-8d0c-f7f9eaf05b6f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.523144] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52072911-1a6b-046a-8d0c-f7f9eaf05b6f, 'name': SearchDatastore_Task, 'duration_secs': 0.010031} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.526724] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1222.527101] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 9f6264eb-2632-4210-b5e8-a58c875a3a85/9f6264eb-2632-4210-b5e8-a58c875a3a85.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1222.527560] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7799f996-7c2a-4d0a-9a1d-30a7ff808cc9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.534640] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1222.534640] env[61906]: value = "task-1333853" [ 1222.534640] env[61906]: _type = "Task" [ 1222.534640] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.545482] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333853, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.043535] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333853, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.472392} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.043969] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9502563-1f96-46a5-b87e-12c0e9375e6e/a9502563-1f96-46a5-b87e-12c0e9375e6e.vmdk to [datastore2] 9f6264eb-2632-4210-b5e8-a58c875a3a85/9f6264eb-2632-4210-b5e8-a58c875a3a85.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1223.044107] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1223.044359] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-19e98111-be10-488b-baae-71134086378e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.050140] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1223.050140] env[61906]: value = "task-1333854" [ 1223.050140] env[61906]: _type = "Task" [ 1223.050140] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.057921] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333854, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.560485] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333854, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061415} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.560768] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1223.561546] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bebe4696-1be1-43ae-8770-a21253e346e1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.583496] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 9f6264eb-2632-4210-b5e8-a58c875a3a85/9f6264eb-2632-4210-b5e8-a58c875a3a85.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1223.583792] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-648e4408-21e5-4492-9041-8a0d2ff5abbb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.607067] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1223.607067] env[61906]: value = "task-1333855" [ 1223.607067] env[61906]: _type = "Task" [ 1223.607067] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.613951] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333855, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.119529] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333855, 'name': ReconfigVM_Task, 'duration_secs': 0.261209} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.119946] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 9f6264eb-2632-4210-b5e8-a58c875a3a85/9f6264eb-2632-4210-b5e8-a58c875a3a85.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1224.120790] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e65ff628-3b99-4f5e-959b-e4d95f53d0fb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.128579] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1224.128579] env[61906]: value = "task-1333856" [ 1224.128579] env[61906]: _type = "Task" [ 1224.128579] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.141940] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333856, 'name': Rename_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.638826] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333856, 'name': Rename_Task, 'duration_secs': 0.127413} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.639122] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1224.639373] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fe3e1a74-5a29-47cf-8542-4f8153a39b03 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.646093] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1224.646093] env[61906]: value = "task-1333857" [ 1224.646093] env[61906]: _type = "Task" [ 1224.646093] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.653126] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333857, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.156368] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333857, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.657530] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333857, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.158183] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333857, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.660040] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333857, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.159673] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333857, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.660590] env[61906]: DEBUG oslo_vmware.api [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333857, 'name': PowerOnVM_Task, 'duration_secs': 2.786286} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.660872] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1227.661152] env[61906]: INFO nova.compute.manager [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Took 9.21 seconds to spawn the instance on the hypervisor. [ 1227.661334] env[61906]: DEBUG nova.compute.manager [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1227.662119] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e53058c3-d3f1-483d-9cad-8a52c608e53b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.180791] env[61906]: INFO nova.compute.manager [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Took 14.16 seconds to build instance. [ 1228.380754] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1228.384390] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1228.384963] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Starting heal instance info cache {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1228.602848] env[61906]: DEBUG nova.compute.manager [req-20be1220-20c8-4311-9b92-6540c574eda3 req-c4e3b907-23b0-4b6e-9012-10dadb34d6fd service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Received event network-changed-0bc154b4-abd3-4632-8b74-c16ded1c94d0 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1228.603085] env[61906]: DEBUG nova.compute.manager [req-20be1220-20c8-4311-9b92-6540c574eda3 req-c4e3b907-23b0-4b6e-9012-10dadb34d6fd service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Refreshing instance network info cache due to event network-changed-0bc154b4-abd3-4632-8b74-c16ded1c94d0. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1228.603732] env[61906]: DEBUG oslo_concurrency.lockutils [req-20be1220-20c8-4311-9b92-6540c574eda3 req-c4e3b907-23b0-4b6e-9012-10dadb34d6fd service nova] Acquiring lock "refresh_cache-9f6264eb-2632-4210-b5e8-a58c875a3a85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1228.603963] env[61906]: DEBUG oslo_concurrency.lockutils [req-20be1220-20c8-4311-9b92-6540c574eda3 req-c4e3b907-23b0-4b6e-9012-10dadb34d6fd service nova] Acquired lock "refresh_cache-9f6264eb-2632-4210-b5e8-a58c875a3a85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1228.604162] env[61906]: DEBUG nova.network.neutron [req-20be1220-20c8-4311-9b92-6540c574eda3 req-c4e3b907-23b0-4b6e-9012-10dadb34d6fd service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Refreshing network info cache for port 0bc154b4-abd3-4632-8b74-c16ded1c94d0 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1228.682688] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b4b19890-1008-4950-8d51-6cce698669cb tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "9f6264eb-2632-4210-b5e8-a58c875a3a85" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.669s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1228.887387] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Didn't find any instances for network info cache update. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1229.303249] env[61906]: DEBUG nova.network.neutron [req-20be1220-20c8-4311-9b92-6540c574eda3 req-c4e3b907-23b0-4b6e-9012-10dadb34d6fd service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Updated VIF entry in instance network info cache for port 0bc154b4-abd3-4632-8b74-c16ded1c94d0. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1229.303629] env[61906]: DEBUG nova.network.neutron [req-20be1220-20c8-4311-9b92-6540c574eda3 req-c4e3b907-23b0-4b6e-9012-10dadb34d6fd service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Updating instance_info_cache with network_info: [{"id": "0bc154b4-abd3-4632-8b74-c16ded1c94d0", "address": "fa:16:3e:2c:d0:74", "network": {"id": "0085e7d2-e26f-4b5d-a282-cf6e0f81786a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1504444833-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.185", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00c03ca0de1447da9d47c83f256efdcd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bc154b4-ab", "ovs_interfaceid": "0bc154b4-abd3-4632-8b74-c16ded1c94d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1229.384866] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1229.385084] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1229.385251] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1229.385405] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1229.806025] env[61906]: DEBUG oslo_concurrency.lockutils [req-20be1220-20c8-4311-9b92-6540c574eda3 req-c4e3b907-23b0-4b6e-9012-10dadb34d6fd service nova] Releasing lock "refresh_cache-9f6264eb-2632-4210-b5e8-a58c875a3a85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1229.887944] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1229.888258] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1229.888444] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1229.888605] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61906) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1229.889546] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96ee80a3-e6e7-4aff-b42f-bddf40368f89 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.897721] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-671ff369-427d-402f-b9be-b25b4e82f096 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.911487] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5b8d738-a25d-4947-955c-5769f888b2fc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.917527] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17e49035-5c12-43cf-b5f3-5b5f205f6965 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.946319] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180375MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=61906) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1229.946462] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1229.946650] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1230.971808] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1230.972079] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 9f6264eb-2632-4210-b5e8-a58c875a3a85 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1230.972174] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1230.972319] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1231.010678] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a6cf935-89b9-49a0-bda7-c19d2d798068 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.018043] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5864300a-cbf6-43fb-bdd7-b89b1c407cef {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.046914] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9e16c27-cf35-4dbb-a379-b39d1992c868 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.053539] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf415c9a-57bc-4785-9338-3f23190c504f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.065884] env[61906]: DEBUG nova.compute.provider_tree [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1231.569052] env[61906]: DEBUG nova.scheduler.client.report [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1232.075191] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1232.075584] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.129s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1233.072049] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1233.576508] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1233.576865] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1233.576865] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1233.577103] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61906) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1253.267539] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1cec179c-6d41-4fb3-8805-1333d7e30a79 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1253.267914] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1cec179c-6d41-4fb3-8805-1333d7e30a79 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1253.268035] env[61906]: DEBUG nova.compute.manager [None req-1cec179c-6d41-4fb3-8805-1333d7e30a79 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1253.268951] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbe227ed-d5b6-4ea8-b970-a924e700625c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.275834] env[61906]: DEBUG nova.compute.manager [None req-1cec179c-6d41-4fb3-8805-1333d7e30a79 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61906) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1253.276477] env[61906]: DEBUG nova.objects.instance [None req-1cec179c-6d41-4fb3-8805-1333d7e30a79 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lazy-loading 'flavor' on Instance uuid 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1253.782822] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cec179c-6d41-4fb3-8805-1333d7e30a79 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1253.783119] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0470a495-3e19-4f6a-98da-3492cfce30cb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.791594] env[61906]: DEBUG oslo_vmware.api [None req-1cec179c-6d41-4fb3-8805-1333d7e30a79 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1253.791594] env[61906]: value = "task-1333858" [ 1253.791594] env[61906]: _type = "Task" [ 1253.791594] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.799811] env[61906]: DEBUG oslo_vmware.api [None req-1cec179c-6d41-4fb3-8805-1333d7e30a79 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333858, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.301073] env[61906]: DEBUG oslo_vmware.api [None req-1cec179c-6d41-4fb3-8805-1333d7e30a79 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333858, 'name': PowerOffVM_Task, 'duration_secs': 0.185556} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1254.301439] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cec179c-6d41-4fb3-8805-1333d7e30a79 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1254.301533] env[61906]: DEBUG nova.compute.manager [None req-1cec179c-6d41-4fb3-8805-1333d7e30a79 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1254.302330] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9d3e8db-ed8a-46bd-8932-fec323161f4d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.813709] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1cec179c-6d41-4fb3-8805-1333d7e30a79 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.546s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1255.662056] env[61906]: DEBUG nova.objects.instance [None req-159cade0-77aa-47a4-ac8c-c3b6ca75feb5 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lazy-loading 'flavor' on Instance uuid 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1256.167544] env[61906]: DEBUG oslo_concurrency.lockutils [None req-159cade0-77aa-47a4-ac8c-c3b6ca75feb5 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "refresh_cache-8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1256.169087] env[61906]: DEBUG oslo_concurrency.lockutils [None req-159cade0-77aa-47a4-ac8c-c3b6ca75feb5 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired lock "refresh_cache-8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1256.169087] env[61906]: DEBUG nova.network.neutron [None req-159cade0-77aa-47a4-ac8c-c3b6ca75feb5 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1256.169087] env[61906]: DEBUG nova.objects.instance [None req-159cade0-77aa-47a4-ac8c-c3b6ca75feb5 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lazy-loading 'info_cache' on Instance uuid 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1256.672469] env[61906]: DEBUG nova.objects.base [None req-159cade0-77aa-47a4-ac8c-c3b6ca75feb5 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Object Instance<8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd> lazy-loaded attributes: flavor,info_cache {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1257.373766] env[61906]: DEBUG nova.network.neutron [None req-159cade0-77aa-47a4-ac8c-c3b6ca75feb5 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Updating instance_info_cache with network_info: [{"id": "dba2c370-204e-45ba-9ce0-793505bd03f5", "address": "fa:16:3e:a9:97:42", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdba2c370-20", "ovs_interfaceid": "dba2c370-204e-45ba-9ce0-793505bd03f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1257.877063] env[61906]: DEBUG oslo_concurrency.lockutils [None req-159cade0-77aa-47a4-ac8c-c3b6ca75feb5 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Releasing lock "refresh_cache-8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1258.380392] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-159cade0-77aa-47a4-ac8c-c3b6ca75feb5 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1258.380718] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cb1ced8d-d674-47e1-9a31-e80cdff1d9d2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.389454] env[61906]: DEBUG oslo_vmware.api [None req-159cade0-77aa-47a4-ac8c-c3b6ca75feb5 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1258.389454] env[61906]: value = "task-1333859" [ 1258.389454] env[61906]: _type = "Task" [ 1258.389454] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.398988] env[61906]: DEBUG oslo_vmware.api [None req-159cade0-77aa-47a4-ac8c-c3b6ca75feb5 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333859, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.900120] env[61906]: DEBUG oslo_vmware.api [None req-159cade0-77aa-47a4-ac8c-c3b6ca75feb5 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333859, 'name': PowerOnVM_Task, 'duration_secs': 0.351513} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.900120] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-159cade0-77aa-47a4-ac8c-c3b6ca75feb5 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1258.900120] env[61906]: DEBUG nova.compute.manager [None req-159cade0-77aa-47a4-ac8c-c3b6ca75feb5 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1258.900651] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-841aed34-ee7b-4c61-aa0b-a2a49d6f9035 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.630378] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d84fb0d3-aea9-4888-8e0a-8fa98e0c345e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.636981] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b50d8db2-588a-4363-b013-3e2863c22c50 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Suspending the VM {{(pid=61906) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1260.637227] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-5e1f14dd-e71b-453d-8c14-4952e0ec41d9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.643362] env[61906]: DEBUG oslo_vmware.api [None req-b50d8db2-588a-4363-b013-3e2863c22c50 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1260.643362] env[61906]: value = "task-1333860" [ 1260.643362] env[61906]: _type = "Task" [ 1260.643362] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.650767] env[61906]: DEBUG oslo_vmware.api [None req-b50d8db2-588a-4363-b013-3e2863c22c50 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333860, 'name': SuspendVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.744710] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "9f6264eb-2632-4210-b5e8-a58c875a3a85" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1260.745134] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "9f6264eb-2632-4210-b5e8-a58c875a3a85" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1260.745379] env[61906]: INFO nova.compute.manager [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Shelving [ 1261.154358] env[61906]: DEBUG oslo_vmware.api [None req-b50d8db2-588a-4363-b013-3e2863c22c50 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333860, 'name': SuspendVM_Task} progress is 75%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.252632] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1261.252919] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0d88e49a-79fc-4b02-8ba2-10476e57dbb3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.259143] env[61906]: DEBUG oslo_vmware.api [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1261.259143] env[61906]: value = "task-1333861" [ 1261.259143] env[61906]: _type = "Task" [ 1261.259143] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.266920] env[61906]: DEBUG oslo_vmware.api [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333861, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.655258] env[61906]: DEBUG oslo_vmware.api [None req-b50d8db2-588a-4363-b013-3e2863c22c50 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333860, 'name': SuspendVM_Task, 'duration_secs': 0.597123} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.655614] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b50d8db2-588a-4363-b013-3e2863c22c50 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Suspended the VM {{(pid=61906) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1261.655667] env[61906]: DEBUG nova.compute.manager [None req-b50d8db2-588a-4363-b013-3e2863c22c50 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1261.656443] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b5c2eb9-1107-42dc-a56b-5669012d97ee {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.768384] env[61906]: DEBUG oslo_vmware.api [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333861, 'name': PowerOffVM_Task, 'duration_secs': 0.203276} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.768667] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1261.769441] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33bc9f9a-9925-4481-912c-3ae5ebd0f849 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.786413] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efe1d3db-dc58-4b1c-b239-617c2074677e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.295939] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Creating Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1262.296311] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-bdfc8d08-e5c0-4d63-8c82-1bf75db3fb3c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.303877] env[61906]: DEBUG oslo_vmware.api [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1262.303877] env[61906]: value = "task-1333862" [ 1262.303877] env[61906]: _type = "Task" [ 1262.303877] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.311689] env[61906]: DEBUG oslo_vmware.api [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333862, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.813756] env[61906]: DEBUG oslo_vmware.api [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333862, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.977674] env[61906]: INFO nova.compute.manager [None req-08a125a5-e4df-4a42-b867-936810f43b08 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Resuming [ 1262.978466] env[61906]: DEBUG nova.objects.instance [None req-08a125a5-e4df-4a42-b867-936810f43b08 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lazy-loading 'flavor' on Instance uuid 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1263.314444] env[61906]: DEBUG oslo_vmware.api [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333862, 'name': CreateSnapshot_Task, 'duration_secs': 0.751253} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.314747] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Created Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1263.315509] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d00e2e8-fc39-4b2c-8646-b54916bfde47 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.832395] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Creating linked-clone VM from snapshot {{(pid=61906) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1263.832706] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-cf0b166b-e3f1-4a33-aba1-7d2fc97b6529 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.841321] env[61906]: DEBUG oslo_vmware.api [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1263.841321] env[61906]: value = "task-1333863" [ 1263.841321] env[61906]: _type = "Task" [ 1263.841321] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1263.849255] env[61906]: DEBUG oslo_vmware.api [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333863, 'name': CloneVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.986735] env[61906]: DEBUG oslo_concurrency.lockutils [None req-08a125a5-e4df-4a42-b867-936810f43b08 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "refresh_cache-8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1263.986951] env[61906]: DEBUG oslo_concurrency.lockutils [None req-08a125a5-e4df-4a42-b867-936810f43b08 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquired lock "refresh_cache-8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1263.987104] env[61906]: DEBUG nova.network.neutron [None req-08a125a5-e4df-4a42-b867-936810f43b08 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1264.352045] env[61906]: DEBUG oslo_vmware.api [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333863, 'name': CloneVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.699550] env[61906]: DEBUG nova.network.neutron [None req-08a125a5-e4df-4a42-b867-936810f43b08 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Updating instance_info_cache with network_info: [{"id": "dba2c370-204e-45ba-9ce0-793505bd03f5", "address": "fa:16:3e:a9:97:42", "network": {"id": "c7c926f2-1f1d-4dd2-862b-17ab45a20efc", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1705417066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d3aa88849f4492aa16ff337b8cfdda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdba2c370-20", "ovs_interfaceid": "dba2c370-204e-45ba-9ce0-793505bd03f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1264.852694] env[61906]: DEBUG oslo_vmware.api [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333863, 'name': CloneVM_Task} progress is 95%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.202624] env[61906]: DEBUG oslo_concurrency.lockutils [None req-08a125a5-e4df-4a42-b867-936810f43b08 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Releasing lock "refresh_cache-8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1265.203612] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0209e056-289b-468e-ac41-f3de110ad7f6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.210137] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-08a125a5-e4df-4a42-b867-936810f43b08 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Resuming the VM {{(pid=61906) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1265.210373] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2d67d3f0-2d3c-45ce-8d76-a19cb5772479 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.216327] env[61906]: DEBUG oslo_vmware.api [None req-08a125a5-e4df-4a42-b867-936810f43b08 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1265.216327] env[61906]: value = "task-1333864" [ 1265.216327] env[61906]: _type = "Task" [ 1265.216327] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.223356] env[61906]: DEBUG oslo_vmware.api [None req-08a125a5-e4df-4a42-b867-936810f43b08 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333864, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.352879] env[61906]: DEBUG oslo_vmware.api [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333863, 'name': CloneVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.727106] env[61906]: DEBUG oslo_vmware.api [None req-08a125a5-e4df-4a42-b867-936810f43b08 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333864, 'name': PowerOnVM_Task, 'duration_secs': 0.488063} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.727456] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-08a125a5-e4df-4a42-b867-936810f43b08 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Resumed the VM {{(pid=61906) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1265.727682] env[61906]: DEBUG nova.compute.manager [None req-08a125a5-e4df-4a42-b867-936810f43b08 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1265.728505] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a95e1c3-f4e9-4f6b-a2cd-f498c1cb4249 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.854903] env[61906]: DEBUG oslo_vmware.api [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333863, 'name': CloneVM_Task, 'duration_secs': 1.539401} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.855336] env[61906]: INFO nova.virt.vmwareapi.vmops [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Created linked-clone VM from snapshot [ 1265.856071] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daa12603-eb46-433b-ae16-07ccca04987b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.863866] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Uploading image 0c4a388b-3efe-41bf-ad3f-178709472a4f {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1265.890098] env[61906]: DEBUG oslo_vmware.rw_handles [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1265.890098] env[61906]: value = "vm-284857" [ 1265.890098] env[61906]: _type = "VirtualMachine" [ 1265.890098] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1265.890440] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-062887b3-78a9-467c-a3f6-2dd5fcdd2a03 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.898086] env[61906]: DEBUG oslo_vmware.rw_handles [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lease: (returnval){ [ 1265.898086] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5287d45e-bec8-af55-b71f-788b8ee94028" [ 1265.898086] env[61906]: _type = "HttpNfcLease" [ 1265.898086] env[61906]: } obtained for exporting VM: (result){ [ 1265.898086] env[61906]: value = "vm-284857" [ 1265.898086] env[61906]: _type = "VirtualMachine" [ 1265.898086] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1265.898355] env[61906]: DEBUG oslo_vmware.api [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the lease: (returnval){ [ 1265.898355] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5287d45e-bec8-af55-b71f-788b8ee94028" [ 1265.898355] env[61906]: _type = "HttpNfcLease" [ 1265.898355] env[61906]: } to be ready. {{(pid=61906) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1265.906127] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1265.906127] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5287d45e-bec8-af55-b71f-788b8ee94028" [ 1265.906127] env[61906]: _type = "HttpNfcLease" [ 1265.906127] env[61906]: } is initializing. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1266.407077] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1266.407077] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5287d45e-bec8-af55-b71f-788b8ee94028" [ 1266.407077] env[61906]: _type = "HttpNfcLease" [ 1266.407077] env[61906]: } is ready. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1266.407415] env[61906]: DEBUG oslo_vmware.rw_handles [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1266.407415] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]5287d45e-bec8-af55-b71f-788b8ee94028" [ 1266.407415] env[61906]: _type = "HttpNfcLease" [ 1266.407415] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1266.408149] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7684e6a9-ffbd-47c9-bbc3-85d1a377fb6e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.415106] env[61906]: DEBUG oslo_vmware.rw_handles [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5251a268-c58e-e919-85c2-9616bfec31cb/disk-0.vmdk from lease info. {{(pid=61906) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1266.415288] env[61906]: DEBUG oslo_vmware.rw_handles [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5251a268-c58e-e919-85c2-9616bfec31cb/disk-0.vmdk for reading. {{(pid=61906) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1266.521906] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e5faaf22-caf0-4154-a188-5f75cd6729b0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.559462] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1266.559788] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1266.560064] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1266.560302] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1266.560514] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1266.562565] env[61906]: INFO nova.compute.manager [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Terminating instance [ 1266.564283] env[61906]: DEBUG nova.compute.manager [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1266.564484] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1266.565294] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b40c4191-9485-4e4f-984c-f0cf5390f2d2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.572181] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1266.572416] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a16a37dd-4e26-49d3-a3f1-0d0101cfa8dc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.578465] env[61906]: DEBUG oslo_vmware.api [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1266.578465] env[61906]: value = "task-1333866" [ 1266.578465] env[61906]: _type = "Task" [ 1266.578465] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.586899] env[61906]: DEBUG oslo_vmware.api [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333866, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.089740] env[61906]: DEBUG oslo_vmware.api [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333866, 'name': PowerOffVM_Task, 'duration_secs': 0.168107} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1267.090103] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1267.090344] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1267.090722] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-29944a18-08a2-41d8-9923-2e8cdbe0124b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.154125] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1267.154568] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1267.154882] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Deleting the datastore file [datastore1] 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1267.155217] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e3bcb57d-2930-4fab-8d1c-de5d86188b50 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.162150] env[61906]: DEBUG oslo_vmware.api [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for the task: (returnval){ [ 1267.162150] env[61906]: value = "task-1333868" [ 1267.162150] env[61906]: _type = "Task" [ 1267.162150] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.170432] env[61906]: DEBUG oslo_vmware.api [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333868, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.672115] env[61906]: DEBUG oslo_vmware.api [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Task: {'id': task-1333868, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.432748} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1267.672450] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1267.672834] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1267.673044] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1267.673354] env[61906]: INFO nova.compute.manager [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1267.673786] env[61906]: DEBUG oslo.service.loopingcall [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1267.674131] env[61906]: DEBUG nova.compute.manager [-] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1267.674270] env[61906]: DEBUG nova.network.neutron [-] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1268.310017] env[61906]: DEBUG nova.compute.manager [req-ef46deb9-6ff5-4423-b571-4c833ff03fee req-5f7ff632-4241-4919-8c13-849727b43da7 service nova] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Received event network-vif-deleted-dba2c370-204e-45ba-9ce0-793505bd03f5 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1268.310548] env[61906]: INFO nova.compute.manager [req-ef46deb9-6ff5-4423-b571-4c833ff03fee req-5f7ff632-4241-4919-8c13-849727b43da7 service nova] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Neutron deleted interface dba2c370-204e-45ba-9ce0-793505bd03f5; detaching it from the instance and deleting it from the info cache [ 1268.310807] env[61906]: DEBUG nova.network.neutron [req-ef46deb9-6ff5-4423-b571-4c833ff03fee req-5f7ff632-4241-4919-8c13-849727b43da7 service nova] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1268.780587] env[61906]: DEBUG nova.network.neutron [-] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1268.813159] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-17373da5-084b-419e-a3e6-f6da4fe949a2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.822824] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86f60bd7-5cb2-4133-90c1-ad59a8d87db2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.846926] env[61906]: DEBUG nova.compute.manager [req-ef46deb9-6ff5-4423-b571-4c833ff03fee req-5f7ff632-4241-4919-8c13-849727b43da7 service nova] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Detach interface failed, port_id=dba2c370-204e-45ba-9ce0-793505bd03f5, reason: Instance 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1269.284235] env[61906]: INFO nova.compute.manager [-] [instance: 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd] Took 1.61 seconds to deallocate network for instance. [ 1269.791226] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1269.791635] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1269.791744] env[61906]: DEBUG nova.objects.instance [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lazy-loading 'resources' on Instance uuid 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1270.343344] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4fb9dde-3ed7-4162-a08e-018288f007d2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.351416] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad09313a-f2ad-4abd-892c-f53cc8ff26af {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.382196] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7704d141-8bb8-4925-a1f8-2b91aa4ad84e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.389629] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3894ff6f-5a01-4eec-b29a-e4f1fbe133c3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.403125] env[61906]: DEBUG nova.compute.provider_tree [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1270.906758] env[61906]: DEBUG nova.scheduler.client.report [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1271.412429] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.621s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1271.434872] env[61906]: INFO nova.scheduler.client.report [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Deleted allocations for instance 8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd [ 1271.945047] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6b81f71-3b9c-4f50-8768-1de2fa87a6c1 tempest-ServerActionsTestJSON-112571910 tempest-ServerActionsTestJSON-112571910-project-member] Lock "8d1bd3d5-6bb8-48b6-bcfe-91061d07cffd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.385s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1273.497969] env[61906]: DEBUG oslo_vmware.rw_handles [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5251a268-c58e-e919-85c2-9616bfec31cb/disk-0.vmdk. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1273.498921] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-088a243e-a32c-46b6-8093-4df023e258c0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.505367] env[61906]: DEBUG oslo_vmware.rw_handles [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5251a268-c58e-e919-85c2-9616bfec31cb/disk-0.vmdk is in state: ready. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1273.505528] env[61906]: ERROR oslo_vmware.rw_handles [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5251a268-c58e-e919-85c2-9616bfec31cb/disk-0.vmdk due to incomplete transfer. [ 1273.505738] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-8b4cc5cb-97a2-4e1c-88cd-34286e8f7543 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.512125] env[61906]: DEBUG oslo_vmware.rw_handles [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5251a268-c58e-e919-85c2-9616bfec31cb/disk-0.vmdk. {{(pid=61906) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1273.512294] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Uploaded image 0c4a388b-3efe-41bf-ad3f-178709472a4f to the Glance image server {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1273.514680] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Destroying the VM {{(pid=61906) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1273.514909] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-60cab7a2-c4de-4fdf-a782-5dc1e31f65fb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.520140] env[61906]: DEBUG oslo_vmware.api [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1273.520140] env[61906]: value = "task-1333869" [ 1273.520140] env[61906]: _type = "Task" [ 1273.520140] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1273.527311] env[61906]: DEBUG oslo_vmware.api [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333869, 'name': Destroy_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1274.029257] env[61906]: DEBUG oslo_vmware.api [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333869, 'name': Destroy_Task, 'duration_secs': 0.284873} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1274.029531] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Destroyed the VM [ 1274.029783] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Deleting Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1274.030048] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-cac5d40d-945d-48b0-8a53-3f56b326bfe1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.037482] env[61906]: DEBUG oslo_vmware.api [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1274.037482] env[61906]: value = "task-1333870" [ 1274.037482] env[61906]: _type = "Task" [ 1274.037482] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1274.045209] env[61906]: DEBUG oslo_vmware.api [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333870, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1274.548854] env[61906]: DEBUG oslo_vmware.api [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333870, 'name': RemoveSnapshot_Task, 'duration_secs': 0.351787} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1274.549189] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Deleted Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1274.549475] env[61906]: DEBUG nova.compute.manager [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1274.550323] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17adf9b6-b986-4f10-b641-344b4643c59e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.063350] env[61906]: INFO nova.compute.manager [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Shelve offloading [ 1275.065028] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1275.065284] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5e4fa327-5a05-434e-a197-033ed12b5e30 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.073316] env[61906]: DEBUG oslo_vmware.api [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1275.073316] env[61906]: value = "task-1333871" [ 1275.073316] env[61906]: _type = "Task" [ 1275.073316] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1275.081922] env[61906]: DEBUG oslo_vmware.api [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333871, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1275.585200] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] VM already powered off {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1275.585721] env[61906]: DEBUG nova.compute.manager [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1275.586302] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b00401bd-79f3-4e8a-a097-c02b3e7ec16d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.592808] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "refresh_cache-9f6264eb-2632-4210-b5e8-a58c875a3a85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1275.592982] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquired lock "refresh_cache-9f6264eb-2632-4210-b5e8-a58c875a3a85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1275.593087] env[61906]: DEBUG nova.network.neutron [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1276.312772] env[61906]: DEBUG nova.network.neutron [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Updating instance_info_cache with network_info: [{"id": "0bc154b4-abd3-4632-8b74-c16ded1c94d0", "address": "fa:16:3e:2c:d0:74", "network": {"id": "0085e7d2-e26f-4b5d-a282-cf6e0f81786a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1504444833-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.185", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00c03ca0de1447da9d47c83f256efdcd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bc154b4-ab", "ovs_interfaceid": "0bc154b4-abd3-4632-8b74-c16ded1c94d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1276.816183] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Releasing lock "refresh_cache-9f6264eb-2632-4210-b5e8-a58c875a3a85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1277.023011] env[61906]: DEBUG nova.compute.manager [req-e93d75e6-1613-4dd3-8c68-6fbe491f4fe5 req-92e6e28a-1930-4f76-8e74-e1ef58a80c11 service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Received event network-vif-unplugged-0bc154b4-abd3-4632-8b74-c16ded1c94d0 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1277.023238] env[61906]: DEBUG oslo_concurrency.lockutils [req-e93d75e6-1613-4dd3-8c68-6fbe491f4fe5 req-92e6e28a-1930-4f76-8e74-e1ef58a80c11 service nova] Acquiring lock "9f6264eb-2632-4210-b5e8-a58c875a3a85-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1277.023453] env[61906]: DEBUG oslo_concurrency.lockutils [req-e93d75e6-1613-4dd3-8c68-6fbe491f4fe5 req-92e6e28a-1930-4f76-8e74-e1ef58a80c11 service nova] Lock "9f6264eb-2632-4210-b5e8-a58c875a3a85-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1277.023625] env[61906]: DEBUG oslo_concurrency.lockutils [req-e93d75e6-1613-4dd3-8c68-6fbe491f4fe5 req-92e6e28a-1930-4f76-8e74-e1ef58a80c11 service nova] Lock "9f6264eb-2632-4210-b5e8-a58c875a3a85-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1277.023800] env[61906]: DEBUG nova.compute.manager [req-e93d75e6-1613-4dd3-8c68-6fbe491f4fe5 req-92e6e28a-1930-4f76-8e74-e1ef58a80c11 service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] No waiting events found dispatching network-vif-unplugged-0bc154b4-abd3-4632-8b74-c16ded1c94d0 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1277.023974] env[61906]: WARNING nova.compute.manager [req-e93d75e6-1613-4dd3-8c68-6fbe491f4fe5 req-92e6e28a-1930-4f76-8e74-e1ef58a80c11 service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Received unexpected event network-vif-unplugged-0bc154b4-abd3-4632-8b74-c16ded1c94d0 for instance with vm_state shelved and task_state shelving_offloading. [ 1277.113195] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1277.114242] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3ae52af-76d9-4cb7-b49b-f556b44af2f8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.121522] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1277.121779] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-94eaf48e-520d-4c9f-aa29-64a1a1679710 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.178039] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1277.178286] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1277.178477] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Deleting the datastore file [datastore2] 9f6264eb-2632-4210-b5e8-a58c875a3a85 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1277.178748] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6746a40c-fbeb-43af-8ea9-b05f38589be6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.187759] env[61906]: DEBUG oslo_vmware.api [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1277.187759] env[61906]: value = "task-1333873" [ 1277.187759] env[61906]: _type = "Task" [ 1277.187759] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1277.194628] env[61906]: DEBUG oslo_vmware.api [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333873, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.697822] env[61906]: DEBUG oslo_vmware.api [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333873, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.121662} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1277.698111] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1277.698270] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1277.698458] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1277.722501] env[61906]: INFO nova.scheduler.client.report [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Deleted allocations for instance 9f6264eb-2632-4210-b5e8-a58c875a3a85 [ 1278.227605] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1278.227924] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1278.228172] env[61906]: DEBUG nova.objects.instance [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lazy-loading 'resources' on Instance uuid 9f6264eb-2632-4210-b5e8-a58c875a3a85 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1278.731102] env[61906]: DEBUG nova.objects.instance [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lazy-loading 'numa_topology' on Instance uuid 9f6264eb-2632-4210-b5e8-a58c875a3a85 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1279.047854] env[61906]: DEBUG nova.compute.manager [req-214371cd-0026-4b9f-a314-db89359897ff req-18196c4b-f76a-4c82-b6cf-8ff4ccfc967e service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Received event network-changed-0bc154b4-abd3-4632-8b74-c16ded1c94d0 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1279.048140] env[61906]: DEBUG nova.compute.manager [req-214371cd-0026-4b9f-a314-db89359897ff req-18196c4b-f76a-4c82-b6cf-8ff4ccfc967e service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Refreshing instance network info cache due to event network-changed-0bc154b4-abd3-4632-8b74-c16ded1c94d0. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1279.048293] env[61906]: DEBUG oslo_concurrency.lockutils [req-214371cd-0026-4b9f-a314-db89359897ff req-18196c4b-f76a-4c82-b6cf-8ff4ccfc967e service nova] Acquiring lock "refresh_cache-9f6264eb-2632-4210-b5e8-a58c875a3a85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1279.048442] env[61906]: DEBUG oslo_concurrency.lockutils [req-214371cd-0026-4b9f-a314-db89359897ff req-18196c4b-f76a-4c82-b6cf-8ff4ccfc967e service nova] Acquired lock "refresh_cache-9f6264eb-2632-4210-b5e8-a58c875a3a85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1279.048608] env[61906]: DEBUG nova.network.neutron [req-214371cd-0026-4b9f-a314-db89359897ff req-18196c4b-f76a-4c82-b6cf-8ff4ccfc967e service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Refreshing network info cache for port 0bc154b4-abd3-4632-8b74-c16ded1c94d0 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1279.234315] env[61906]: DEBUG nova.objects.base [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Object Instance<9f6264eb-2632-4210-b5e8-a58c875a3a85> lazy-loaded attributes: resources,numa_topology {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1279.250071] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0deae91b-507c-4452-922c-899f060b254e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.256847] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c44c0f72-189f-4767-8ccf-0cec95b4c647 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.287257] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66843384-5cf1-41c0-be69-50d4b2fd40f8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.294185] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75e537ae-bc5d-473a-9aee-213f18635fdc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.307810] env[61906]: DEBUG nova.compute.provider_tree [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1279.752170] env[61906]: DEBUG nova.network.neutron [req-214371cd-0026-4b9f-a314-db89359897ff req-18196c4b-f76a-4c82-b6cf-8ff4ccfc967e service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Updated VIF entry in instance network info cache for port 0bc154b4-abd3-4632-8b74-c16ded1c94d0. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1279.752538] env[61906]: DEBUG nova.network.neutron [req-214371cd-0026-4b9f-a314-db89359897ff req-18196c4b-f76a-4c82-b6cf-8ff4ccfc967e service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Updating instance_info_cache with network_info: [{"id": "0bc154b4-abd3-4632-8b74-c16ded1c94d0", "address": "fa:16:3e:2c:d0:74", "network": {"id": "0085e7d2-e26f-4b5d-a282-cf6e0f81786a", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1504444833-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.185", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00c03ca0de1447da9d47c83f256efdcd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap0bc154b4-ab", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1279.810309] env[61906]: DEBUG nova.scheduler.client.report [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1280.255211] env[61906]: DEBUG oslo_concurrency.lockutils [req-214371cd-0026-4b9f-a314-db89359897ff req-18196c4b-f76a-4c82-b6cf-8ff4ccfc967e service nova] Releasing lock "refresh_cache-9f6264eb-2632-4210-b5e8-a58c875a3a85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1280.303022] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "9f6264eb-2632-4210-b5e8-a58c875a3a85" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1280.314911] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.087s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1280.822603] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6956e112-9414-4e28-89a5-c1c3cfcfdf92 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "9f6264eb-2632-4210-b5e8-a58c875a3a85" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 20.077s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1280.823386] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "9f6264eb-2632-4210-b5e8-a58c875a3a85" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.521s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1280.823573] env[61906]: INFO nova.compute.manager [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Unshelving [ 1281.847413] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1281.847721] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1281.847861] env[61906]: DEBUG nova.objects.instance [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lazy-loading 'pci_requests' on Instance uuid 9f6264eb-2632-4210-b5e8-a58c875a3a85 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1282.351908] env[61906]: DEBUG nova.objects.instance [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lazy-loading 'numa_topology' on Instance uuid 9f6264eb-2632-4210-b5e8-a58c875a3a85 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1282.854591] env[61906]: INFO nova.compute.claims [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1283.890366] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc964eef-7d15-4cdc-b98e-43a435ac918e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.897434] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91fe3b21-a736-453a-8343-f4e70eb9bb95 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.925784] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e27d60e4-4fd8-40a9-b581-f260589c9e62 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.932311] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beeee633-14c2-4e43-9ec6-c9e8a73ee9fb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.944570] env[61906]: DEBUG nova.compute.provider_tree [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1284.448077] env[61906]: DEBUG nova.scheduler.client.report [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1284.954727] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.107s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1284.983297] env[61906]: INFO nova.network.neutron [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Updating port 0bc154b4-abd3-4632-8b74-c16ded1c94d0 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1286.348786] env[61906]: DEBUG nova.compute.manager [req-87c368c1-7f67-45d4-9fbf-f7a2db0c2f99 req-ea3f80de-cad0-4637-afff-21d0dabde477 service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Received event network-vif-plugged-0bc154b4-abd3-4632-8b74-c16ded1c94d0 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1286.349190] env[61906]: DEBUG oslo_concurrency.lockutils [req-87c368c1-7f67-45d4-9fbf-f7a2db0c2f99 req-ea3f80de-cad0-4637-afff-21d0dabde477 service nova] Acquiring lock "9f6264eb-2632-4210-b5e8-a58c875a3a85-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1286.349258] env[61906]: DEBUG oslo_concurrency.lockutils [req-87c368c1-7f67-45d4-9fbf-f7a2db0c2f99 req-ea3f80de-cad0-4637-afff-21d0dabde477 service nova] Lock "9f6264eb-2632-4210-b5e8-a58c875a3a85-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1286.349565] env[61906]: DEBUG oslo_concurrency.lockutils [req-87c368c1-7f67-45d4-9fbf-f7a2db0c2f99 req-ea3f80de-cad0-4637-afff-21d0dabde477 service nova] Lock "9f6264eb-2632-4210-b5e8-a58c875a3a85-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1286.349828] env[61906]: DEBUG nova.compute.manager [req-87c368c1-7f67-45d4-9fbf-f7a2db0c2f99 req-ea3f80de-cad0-4637-afff-21d0dabde477 service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] No waiting events found dispatching network-vif-plugged-0bc154b4-abd3-4632-8b74-c16ded1c94d0 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1286.350025] env[61906]: WARNING nova.compute.manager [req-87c368c1-7f67-45d4-9fbf-f7a2db0c2f99 req-ea3f80de-cad0-4637-afff-21d0dabde477 service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Received unexpected event network-vif-plugged-0bc154b4-abd3-4632-8b74-c16ded1c94d0 for instance with vm_state shelved_offloaded and task_state spawning. [ 1286.433175] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "refresh_cache-9f6264eb-2632-4210-b5e8-a58c875a3a85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1286.433383] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquired lock "refresh_cache-9f6264eb-2632-4210-b5e8-a58c875a3a85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1286.433600] env[61906]: DEBUG nova.network.neutron [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1287.186135] env[61906]: DEBUG nova.network.neutron [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Updating instance_info_cache with network_info: [{"id": "0bc154b4-abd3-4632-8b74-c16ded1c94d0", "address": "fa:16:3e:2c:d0:74", "network": {"id": "0085e7d2-e26f-4b5d-a282-cf6e0f81786a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1504444833-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.185", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00c03ca0de1447da9d47c83f256efdcd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bc154b4-ab", "ovs_interfaceid": "0bc154b4-abd3-4632-8b74-c16ded1c94d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1287.688627] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Releasing lock "refresh_cache-9f6264eb-2632-4210-b5e8-a58c875a3a85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1287.714955] env[61906]: DEBUG nova.virt.hardware [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T06:18:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='14319c9283a8b0dc1c684f244fc1adcc',container_format='bare',created_at=2024-10-22T06:31:52Z,direct_url=,disk_format='vmdk',id=0c4a388b-3efe-41bf-ad3f-178709472a4f,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-715948020-shelved',owner='00c03ca0de1447da9d47c83f256efdcd',properties=ImageMetaProps,protected=,size=31662592,status='active',tags=,updated_at=2024-10-22T06:32:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1287.715238] env[61906]: DEBUG nova.virt.hardware [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1287.715401] env[61906]: DEBUG nova.virt.hardware [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1287.715589] env[61906]: DEBUG nova.virt.hardware [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1287.715745] env[61906]: DEBUG nova.virt.hardware [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1287.715894] env[61906]: DEBUG nova.virt.hardware [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1287.716140] env[61906]: DEBUG nova.virt.hardware [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1287.716306] env[61906]: DEBUG nova.virt.hardware [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1287.716508] env[61906]: DEBUG nova.virt.hardware [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1287.716675] env[61906]: DEBUG nova.virt.hardware [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1287.716853] env[61906]: DEBUG nova.virt.hardware [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1287.717714] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3538e770-5508-4957-8e1f-00f217f098e7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.725789] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3bb1440-1a36-49be-8af5-010dafdc217e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.738603] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2c:d0:74', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c6934071-bf85-4591-9c7d-55c7ea131262', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0bc154b4-abd3-4632-8b74-c16ded1c94d0', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1287.745856] env[61906]: DEBUG oslo.service.loopingcall [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1287.746097] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1287.746300] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ed539fb5-3713-4f22-982b-e2954a7785da {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.764121] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1287.764121] env[61906]: value = "task-1333874" [ 1287.764121] env[61906]: _type = "Task" [ 1287.764121] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.771101] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333874, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.273537] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333874, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.372386] env[61906]: DEBUG nova.compute.manager [req-9222e513-e75d-457c-a3f1-ed7042bf613e req-f1707e92-fdd8-4540-b767-042cfa50b87a service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Received event network-changed-0bc154b4-abd3-4632-8b74-c16ded1c94d0 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1288.372589] env[61906]: DEBUG nova.compute.manager [req-9222e513-e75d-457c-a3f1-ed7042bf613e req-f1707e92-fdd8-4540-b767-042cfa50b87a service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Refreshing instance network info cache due to event network-changed-0bc154b4-abd3-4632-8b74-c16ded1c94d0. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1288.372812] env[61906]: DEBUG oslo_concurrency.lockutils [req-9222e513-e75d-457c-a3f1-ed7042bf613e req-f1707e92-fdd8-4540-b767-042cfa50b87a service nova] Acquiring lock "refresh_cache-9f6264eb-2632-4210-b5e8-a58c875a3a85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1288.372958] env[61906]: DEBUG oslo_concurrency.lockutils [req-9222e513-e75d-457c-a3f1-ed7042bf613e req-f1707e92-fdd8-4540-b767-042cfa50b87a service nova] Acquired lock "refresh_cache-9f6264eb-2632-4210-b5e8-a58c875a3a85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1288.373136] env[61906]: DEBUG nova.network.neutron [req-9222e513-e75d-457c-a3f1-ed7042bf613e req-f1707e92-fdd8-4540-b767-042cfa50b87a service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Refreshing network info cache for port 0bc154b4-abd3-4632-8b74-c16ded1c94d0 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1288.775205] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333874, 'name': CreateVM_Task, 'duration_secs': 0.596259} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1288.775589] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1288.776022] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0c4a388b-3efe-41bf-ad3f-178709472a4f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1288.776199] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0c4a388b-3efe-41bf-ad3f-178709472a4f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1288.776646] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0c4a388b-3efe-41bf-ad3f-178709472a4f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1288.776911] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d686cf61-ba51-44cf-b6b7-48dc626de1de {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.781218] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1288.781218] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52515fd7-d343-dcdd-5425-a211361a54f5" [ 1288.781218] env[61906]: _type = "Task" [ 1288.781218] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1288.788311] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52515fd7-d343-dcdd-5425-a211361a54f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.886295] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1289.061977] env[61906]: DEBUG nova.network.neutron [req-9222e513-e75d-457c-a3f1-ed7042bf613e req-f1707e92-fdd8-4540-b767-042cfa50b87a service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Updated VIF entry in instance network info cache for port 0bc154b4-abd3-4632-8b74-c16ded1c94d0. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1289.062351] env[61906]: DEBUG nova.network.neutron [req-9222e513-e75d-457c-a3f1-ed7042bf613e req-f1707e92-fdd8-4540-b767-042cfa50b87a service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Updating instance_info_cache with network_info: [{"id": "0bc154b4-abd3-4632-8b74-c16ded1c94d0", "address": "fa:16:3e:2c:d0:74", "network": {"id": "0085e7d2-e26f-4b5d-a282-cf6e0f81786a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1504444833-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.185", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00c03ca0de1447da9d47c83f256efdcd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bc154b4-ab", "ovs_interfaceid": "0bc154b4-abd3-4632-8b74-c16ded1c94d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1289.291092] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0c4a388b-3efe-41bf-ad3f-178709472a4f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1289.291381] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Processing image 0c4a388b-3efe-41bf-ad3f-178709472a4f {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1289.291627] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0c4a388b-3efe-41bf-ad3f-178709472a4f/0c4a388b-3efe-41bf-ad3f-178709472a4f.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1289.291788] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0c4a388b-3efe-41bf-ad3f-178709472a4f/0c4a388b-3efe-41bf-ad3f-178709472a4f.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1289.291994] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1289.292266] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-426a19c8-22f5-4169-84f4-08008358c90a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.307824] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1289.308014] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1289.308714] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56f76859-ce31-4432-b927-82e1b00b69d0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.313565] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1289.313565] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52d88e41-07f9-6cc9-e314-d8dbfd6bdd29" [ 1289.313565] env[61906]: _type = "Task" [ 1289.313565] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1289.320637] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': session[527d9377-231f-9983-7520-ff5d4903bb51]52d88e41-07f9-6cc9-e314-d8dbfd6bdd29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1289.384654] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1289.565374] env[61906]: DEBUG oslo_concurrency.lockutils [req-9222e513-e75d-457c-a3f1-ed7042bf613e req-f1707e92-fdd8-4540-b767-042cfa50b87a service nova] Releasing lock "refresh_cache-9f6264eb-2632-4210-b5e8-a58c875a3a85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1289.824149] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Preparing fetch location {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1289.824564] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Fetch image to [datastore2] OSTACK_IMG_ed41a5d8-df7a-49ba-bcc0-be3482f2260d/OSTACK_IMG_ed41a5d8-df7a-49ba-bcc0-be3482f2260d.vmdk {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1289.824683] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Downloading stream optimized image 0c4a388b-3efe-41bf-ad3f-178709472a4f to [datastore2] OSTACK_IMG_ed41a5d8-df7a-49ba-bcc0-be3482f2260d/OSTACK_IMG_ed41a5d8-df7a-49ba-bcc0-be3482f2260d.vmdk on the data store datastore2 as vApp {{(pid=61906) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1289.824867] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Downloading image file data 0c4a388b-3efe-41bf-ad3f-178709472a4f to the ESX as VM named 'OSTACK_IMG_ed41a5d8-df7a-49ba-bcc0-be3482f2260d' {{(pid=61906) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1289.887983] env[61906]: DEBUG oslo_vmware.rw_handles [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1289.887983] env[61906]: value = "resgroup-9" [ 1289.887983] env[61906]: _type = "ResourcePool" [ 1289.887983] env[61906]: }. {{(pid=61906) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1289.888311] env[61906]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-c1f45759-ddd2-4cf2-9ae5-a25cbe641f81 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.907550] env[61906]: DEBUG oslo_vmware.rw_handles [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lease: (returnval){ [ 1289.907550] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52a18881-b84e-2078-8b75-c16fd55f8adf" [ 1289.907550] env[61906]: _type = "HttpNfcLease" [ 1289.907550] env[61906]: } obtained for vApp import into resource pool (val){ [ 1289.907550] env[61906]: value = "resgroup-9" [ 1289.907550] env[61906]: _type = "ResourcePool" [ 1289.907550] env[61906]: }. {{(pid=61906) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1289.907832] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the lease: (returnval){ [ 1289.907832] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52a18881-b84e-2078-8b75-c16fd55f8adf" [ 1289.907832] env[61906]: _type = "HttpNfcLease" [ 1289.907832] env[61906]: } to be ready. {{(pid=61906) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1289.913465] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1289.913465] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52a18881-b84e-2078-8b75-c16fd55f8adf" [ 1289.913465] env[61906]: _type = "HttpNfcLease" [ 1289.913465] env[61906]: } is initializing. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1290.385159] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1290.385344] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Starting heal instance info cache {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1290.385466] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Rebuilding the list of instances to heal {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1290.415102] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1290.415102] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52a18881-b84e-2078-8b75-c16fd55f8adf" [ 1290.415102] env[61906]: _type = "HttpNfcLease" [ 1290.415102] env[61906]: } is ready. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1290.416050] env[61906]: DEBUG oslo_vmware.rw_handles [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1290.416050] env[61906]: value = "session[527d9377-231f-9983-7520-ff5d4903bb51]52a18881-b84e-2078-8b75-c16fd55f8adf" [ 1290.416050] env[61906]: _type = "HttpNfcLease" [ 1290.416050] env[61906]: }. {{(pid=61906) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1290.416312] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee0e5907-881f-444b-ba54-a974cea73d7b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.422864] env[61906]: DEBUG oslo_vmware.rw_handles [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5297a55c-8bc2-50b1-4ec3-5753b0b1eeb6/disk-0.vmdk from lease info. {{(pid=61906) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1290.423061] env[61906]: DEBUG oslo_vmware.rw_handles [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Creating HTTP connection to write to file with size = 31662592 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5297a55c-8bc2-50b1-4ec3-5753b0b1eeb6/disk-0.vmdk. {{(pid=61906) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1290.485572] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-021018b9-def9-4fe4-978d-f55785a30af7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.890116] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "refresh_cache-9f6264eb-2632-4210-b5e8-a58c875a3a85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1290.890534] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquired lock "refresh_cache-9f6264eb-2632-4210-b5e8-a58c875a3a85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1290.890534] env[61906]: DEBUG nova.network.neutron [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Forcefully refreshing network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1290.890683] env[61906]: DEBUG nova.objects.instance [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lazy-loading 'info_cache' on Instance uuid 9f6264eb-2632-4210-b5e8-a58c875a3a85 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1291.504325] env[61906]: DEBUG oslo_vmware.rw_handles [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Completed reading data from the image iterator. {{(pid=61906) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1291.504516] env[61906]: DEBUG oslo_vmware.rw_handles [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5297a55c-8bc2-50b1-4ec3-5753b0b1eeb6/disk-0.vmdk. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1291.505435] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-420e7ecd-4d29-4515-a495-3e3e6a6af24b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.512783] env[61906]: DEBUG oslo_vmware.rw_handles [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5297a55c-8bc2-50b1-4ec3-5753b0b1eeb6/disk-0.vmdk is in state: ready. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1291.512979] env[61906]: DEBUG oslo_vmware.rw_handles [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5297a55c-8bc2-50b1-4ec3-5753b0b1eeb6/disk-0.vmdk. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1291.513198] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-ee4d9bfa-361f-40fb-9ae4-55464a961150 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.699651] env[61906]: DEBUG oslo_vmware.rw_handles [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5297a55c-8bc2-50b1-4ec3-5753b0b1eeb6/disk-0.vmdk. {{(pid=61906) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1291.699931] env[61906]: INFO nova.virt.vmwareapi.images [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Downloaded image file data 0c4a388b-3efe-41bf-ad3f-178709472a4f [ 1291.700769] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3bada3a-37ee-4f46-8d91-63381e8cc826 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.716645] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-113f083d-c8f5-496d-afae-4ad791180ec3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.736866] env[61906]: INFO nova.virt.vmwareapi.images [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] The imported VM was unregistered [ 1291.739244] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Caching image {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1291.739477] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Creating directory with path [datastore2] devstack-image-cache_base/0c4a388b-3efe-41bf-ad3f-178709472a4f {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1291.739725] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fc532c0f-71a9-43b5-9e1b-7ec72cf317c1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.748354] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Created directory with path [datastore2] devstack-image-cache_base/0c4a388b-3efe-41bf-ad3f-178709472a4f {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1291.748650] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_ed41a5d8-df7a-49ba-bcc0-be3482f2260d/OSTACK_IMG_ed41a5d8-df7a-49ba-bcc0-be3482f2260d.vmdk to [datastore2] devstack-image-cache_base/0c4a388b-3efe-41bf-ad3f-178709472a4f/0c4a388b-3efe-41bf-ad3f-178709472a4f.vmdk. {{(pid=61906) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1291.748923] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-821570d7-7143-4359-b798-cd972d16c2c1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.754847] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1291.754847] env[61906]: value = "task-1333877" [ 1291.754847] env[61906]: _type = "Task" [ 1291.754847] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1291.761649] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333877, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.264124] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333877, 'name': MoveVirtualDisk_Task} progress is 26%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.653194] env[61906]: DEBUG nova.network.neutron [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Updating instance_info_cache with network_info: [{"id": "0bc154b4-abd3-4632-8b74-c16ded1c94d0", "address": "fa:16:3e:2c:d0:74", "network": {"id": "0085e7d2-e26f-4b5d-a282-cf6e0f81786a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1504444833-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.185", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00c03ca0de1447da9d47c83f256efdcd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bc154b4-ab", "ovs_interfaceid": "0bc154b4-abd3-4632-8b74-c16ded1c94d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1292.766238] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333877, 'name': MoveVirtualDisk_Task} progress is 49%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.155796] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Releasing lock "refresh_cache-9f6264eb-2632-4210-b5e8-a58c875a3a85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1293.155972] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Updated the network info_cache for instance {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1293.156060] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1293.156231] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1293.156384] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1293.156581] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1293.156762] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1293.156900] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61906) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1293.157059] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1293.265382] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333877, 'name': MoveVirtualDisk_Task} progress is 74%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.660437] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1293.660720] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1293.660934] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1293.661149] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61906) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1293.662199] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f62b7ef6-2c38-4e68-b889-00650187a3b0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.671028] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db4fcfd-7d14-4c82-9c10-5bef8fe9e4ad {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.686775] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3850d782-7f8c-4876-b2fb-cf5fb7c8d96c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.695206] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-873d4018-4aa6-4e8e-b4f5-405a917fec69 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.726259] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181097MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=61906) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1293.726432] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1293.726662] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1293.765777] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333877, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1294.266316] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333877, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.082314} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1294.266717] env[61906]: INFO nova.virt.vmwareapi.ds_util [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_ed41a5d8-df7a-49ba-bcc0-be3482f2260d/OSTACK_IMG_ed41a5d8-df7a-49ba-bcc0-be3482f2260d.vmdk to [datastore2] devstack-image-cache_base/0c4a388b-3efe-41bf-ad3f-178709472a4f/0c4a388b-3efe-41bf-ad3f-178709472a4f.vmdk. [ 1294.266816] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Cleaning up location [datastore2] OSTACK_IMG_ed41a5d8-df7a-49ba-bcc0-be3482f2260d {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1294.266966] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_ed41a5d8-df7a-49ba-bcc0-be3482f2260d {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1294.267224] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-027ad31c-6107-4ed3-aab1-c849194ec4a5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.273510] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1294.273510] env[61906]: value = "task-1333878" [ 1294.273510] env[61906]: _type = "Task" [ 1294.273510] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1294.280577] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333878, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1294.751460] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Instance 9f6264eb-2632-4210-b5e8-a58c875a3a85 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1294.751667] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1294.751814] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1294.779699] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05f1277f-2e2e-4059-96f2-d060e8d1eee0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.784878] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333878, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.033846} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1294.785453] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1294.785641] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0c4a388b-3efe-41bf-ad3f-178709472a4f/0c4a388b-3efe-41bf-ad3f-178709472a4f.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1294.785911] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0c4a388b-3efe-41bf-ad3f-178709472a4f/0c4a388b-3efe-41bf-ad3f-178709472a4f.vmdk to [datastore2] 9f6264eb-2632-4210-b5e8-a58c875a3a85/9f6264eb-2632-4210-b5e8-a58c875a3a85.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1294.786170] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-14152d09-8d2d-43ab-83fd-01f1a5a32d98 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.790501] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce26d1e6-0f5e-4909-b0a0-708b001074b9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.794137] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1294.794137] env[61906]: value = "task-1333879" [ 1294.794137] env[61906]: _type = "Task" [ 1294.794137] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1294.820087] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ae7564c-fde2-4140-ab13-e2548afb425e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.824748] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333879, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1294.828997] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c6a15e5-db18-471b-8fa5-eda6bfd50f2c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.842079] env[61906]: DEBUG nova.compute.provider_tree [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1295.304680] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333879, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1295.345071] env[61906]: DEBUG nova.scheduler.client.report [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1295.805713] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333879, 'name': CopyVirtualDisk_Task} progress is 49%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1295.850835] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1295.851073] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.124s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1296.305940] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333879, 'name': CopyVirtualDisk_Task} progress is 71%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1296.806982] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333879, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1297.307973] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333879, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.044221} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1297.308368] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0c4a388b-3efe-41bf-ad3f-178709472a4f/0c4a388b-3efe-41bf-ad3f-178709472a4f.vmdk to [datastore2] 9f6264eb-2632-4210-b5e8-a58c875a3a85/9f6264eb-2632-4210-b5e8-a58c875a3a85.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1297.309095] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e64a12fb-f624-4d81-921b-a311be2ffb6f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.330278] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 9f6264eb-2632-4210-b5e8-a58c875a3a85/9f6264eb-2632-4210-b5e8-a58c875a3a85.vmdk or device None with type streamOptimized {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1297.330489] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-43d3ebf6-cc52-4ed3-bb5d-4591732cce2e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.348830] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1297.348830] env[61906]: value = "task-1333880" [ 1297.348830] env[61906]: _type = "Task" [ 1297.348830] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1297.355966] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333880, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1297.861989] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333880, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1298.360824] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333880, 'name': ReconfigVM_Task, 'duration_secs': 0.668776} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1298.361226] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 9f6264eb-2632-4210-b5e8-a58c875a3a85/9f6264eb-2632-4210-b5e8-a58c875a3a85.vmdk or device None with type streamOptimized {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1298.361736] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9fed7839-2edb-4126-9514-09b9c94cd4c1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.367759] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1298.367759] env[61906]: value = "task-1333881" [ 1298.367759] env[61906]: _type = "Task" [ 1298.367759] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1298.377990] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333881, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1298.877272] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333881, 'name': Rename_Task, 'duration_secs': 0.127429} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1298.877616] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1298.877890] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-07f76ccf-7b49-47d3-a1b5-86addba7219b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.883672] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1298.883672] env[61906]: value = "task-1333882" [ 1298.883672] env[61906]: _type = "Task" [ 1298.883672] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1298.891752] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333882, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1300.055651] env[61906]: DEBUG oslo_vmware.api [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333882, 'name': PowerOnVM_Task, 'duration_secs': 0.5016} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1300.056148] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1300.146541] env[61906]: DEBUG nova.compute.manager [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1300.147524] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e0ef8f-237d-4493-b8d1-0a3ef51bf688 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.664083] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3c792495-1dea-4fb4-b650-4ccf86ef4f1a tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "9f6264eb-2632-4210-b5e8-a58c875a3a85" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 19.840s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1336.265542] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "9f6264eb-2632-4210-b5e8-a58c875a3a85" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1336.265973] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "9f6264eb-2632-4210-b5e8-a58c875a3a85" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1336.266039] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "9f6264eb-2632-4210-b5e8-a58c875a3a85-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1336.266208] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "9f6264eb-2632-4210-b5e8-a58c875a3a85-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1336.266390] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "9f6264eb-2632-4210-b5e8-a58c875a3a85-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1336.268828] env[61906]: INFO nova.compute.manager [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Terminating instance [ 1336.270585] env[61906]: DEBUG nova.compute.manager [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1336.270785] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1336.271641] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b325798-d620-4801-a3dd-0b75afa40020 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.279114] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1336.279343] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b823f8a6-e503-4f5a-93c4-72e9a46d92f7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.285112] env[61906]: DEBUG oslo_vmware.api [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1336.285112] env[61906]: value = "task-1333883" [ 1336.285112] env[61906]: _type = "Task" [ 1336.285112] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1336.292061] env[61906]: DEBUG oslo_vmware.api [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333883, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1336.795088] env[61906]: DEBUG oslo_vmware.api [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333883, 'name': PowerOffVM_Task, 'duration_secs': 0.159492} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1336.795393] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1336.795567] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1336.795819] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cb6b5bfa-5331-4bd1-99e4-321ac43bcd87 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.854460] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1336.854716] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1336.854913] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Deleting the datastore file [datastore2] 9f6264eb-2632-4210-b5e8-a58c875a3a85 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1336.855214] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aa45c2fd-fe71-4c8f-add0-f6892af9baed {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.861298] env[61906]: DEBUG oslo_vmware.api [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for the task: (returnval){ [ 1336.861298] env[61906]: value = "task-1333885" [ 1336.861298] env[61906]: _type = "Task" [ 1336.861298] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1336.868332] env[61906]: DEBUG oslo_vmware.api [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333885, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1337.371234] env[61906]: DEBUG oslo_vmware.api [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Task: {'id': task-1333885, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.123286} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1337.371713] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1337.371713] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1337.371826] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1337.372044] env[61906]: INFO nova.compute.manager [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1337.372258] env[61906]: DEBUG oslo.service.loopingcall [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1337.372479] env[61906]: DEBUG nova.compute.manager [-] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1337.372575] env[61906]: DEBUG nova.network.neutron [-] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1337.784407] env[61906]: DEBUG nova.compute.manager [req-7780eb4b-99db-47a8-b227-ce57011fe1fb req-fe327394-6321-4fad-9027-e6c2d633a9a4 service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Received event network-vif-deleted-0bc154b4-abd3-4632-8b74-c16ded1c94d0 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1337.784615] env[61906]: INFO nova.compute.manager [req-7780eb4b-99db-47a8-b227-ce57011fe1fb req-fe327394-6321-4fad-9027-e6c2d633a9a4 service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Neutron deleted interface 0bc154b4-abd3-4632-8b74-c16ded1c94d0; detaching it from the instance and deleting it from the info cache [ 1337.784781] env[61906]: DEBUG nova.network.neutron [req-7780eb4b-99db-47a8-b227-ce57011fe1fb req-fe327394-6321-4fad-9027-e6c2d633a9a4 service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1338.265149] env[61906]: DEBUG nova.network.neutron [-] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1338.288042] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8197c9aa-92bb-4c3f-afae-1f9f93a3656a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.297065] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ec82d6-6635-4399-a4b3-4632aef484a3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.320012] env[61906]: DEBUG nova.compute.manager [req-7780eb4b-99db-47a8-b227-ce57011fe1fb req-fe327394-6321-4fad-9027-e6c2d633a9a4 service nova] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Detach interface failed, port_id=0bc154b4-abd3-4632-8b74-c16ded1c94d0, reason: Instance 9f6264eb-2632-4210-b5e8-a58c875a3a85 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1338.768295] env[61906]: INFO nova.compute.manager [-] [instance: 9f6264eb-2632-4210-b5e8-a58c875a3a85] Took 1.40 seconds to deallocate network for instance. [ 1339.274548] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1339.274826] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1339.275061] env[61906]: DEBUG nova.objects.instance [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lazy-loading 'resources' on Instance uuid 9f6264eb-2632-4210-b5e8-a58c875a3a85 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1339.795582] env[61906]: DEBUG nova.scheduler.client.report [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Refreshing inventories for resource provider a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1339.809775] env[61906]: DEBUG nova.scheduler.client.report [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Updating ProviderTree inventory for provider a5b57df7-a16e-44f9-9b2d-23c518860263 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1339.809984] env[61906]: DEBUG nova.compute.provider_tree [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Updating inventory in ProviderTree for provider a5b57df7-a16e-44f9-9b2d-23c518860263 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1339.820426] env[61906]: DEBUG nova.scheduler.client.report [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Refreshing aggregate associations for resource provider a5b57df7-a16e-44f9-9b2d-23c518860263, aggregates: None {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1339.837063] env[61906]: DEBUG nova.scheduler.client.report [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Refreshing trait associations for resource provider a5b57df7-a16e-44f9-9b2d-23c518860263, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64 {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1339.861505] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ce0781-0054-44cd-bf75-8d41409c131f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.868809] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b991344-c1b8-4451-a72d-e606619b0944 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.896954] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d388084a-d909-448d-86be-74581c5bf3e6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.903728] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90beebf1-89d5-4991-9e23-f01df23f239f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.917143] env[61906]: DEBUG nova.compute.provider_tree [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1340.420580] env[61906]: DEBUG nova.scheduler.client.report [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1340.925311] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.650s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1340.942896] env[61906]: INFO nova.scheduler.client.report [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Deleted allocations for instance 9f6264eb-2632-4210-b5e8-a58c875a3a85 [ 1341.451551] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c1accc2-8018-4d45-92bf-9b813a034e73 tempest-AttachVolumeShelveTestJSON-1400285163 tempest-AttachVolumeShelveTestJSON-1400285163-project-member] Lock "9f6264eb-2632-4210-b5e8-a58c875a3a85" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.185s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1354.080263] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1354.080700] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1354.080700] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Starting heal instance info cache {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1354.080855] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Rebuilding the list of instances to heal {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1354.583604] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Didn't find any instances for network info cache update. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1354.583896] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1354.583981] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1354.584148] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1354.584300] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1354.584441] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1354.584587] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1354.584714] env[61906]: DEBUG nova.compute.manager [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61906) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1354.584860] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1355.087818] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1355.088242] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1355.088296] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1355.088422] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61906) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1355.089654] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b5f53cb-0eac-417c-982d-dac14118e07d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.097255] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07bf2814-8613-4e57-9ffc-2c5741af2b54 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.110635] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79fee00f-f6f7-4be7-8560-f1fd970340a4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.116485] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3a66687-6122-4e38-88bc-57d96ce0cd0a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.145091] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181421MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=61906) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1355.145237] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1355.145423] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1356.166042] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1356.166281] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1356.178466] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34fc6f62-91fb-48b4-afaf-12e780bbf807 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.185215] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bf1592f-fb90-4141-a3c0-8c9bb04db9fd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.213462] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b45f8ad-ec60-4a29-9132-959e53a3adf3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.219795] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-250def7a-cb53-4c95-8de0-6a7c6ea823e3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.232747] env[61906]: DEBUG nova.compute.provider_tree [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed in ProviderTree for provider: a5b57df7-a16e-44f9-9b2d-23c518860263 {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1356.735957] env[61906]: DEBUG nova.scheduler.client.report [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Inventory has not changed for provider a5b57df7-a16e-44f9-9b2d-23c518860263 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1357.240632] env[61906]: DEBUG nova.compute.resource_tracker [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1357.240986] env[61906]: DEBUG oslo_concurrency.lockutils [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.095s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1357.541300] env[61906]: DEBUG oslo_service.periodic_task [None req-99a04db4-b120-4180-a4c8-dab15e9e6a5f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}}